var/home/core/zuul-output/0000755000175000017500000000000015154752105014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015154764171015504 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000247236215154764106020277 0ustar corecoreFikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gfͅ2|VYI_翪|mvſFެxۻf+ovpZj>?xI[mEy},fۮWe~7Nû/wb~1;ZxsY~ݳ( 2[$7۫j{Zw鶾z?&~|XLXlN_/:oXx$%X"LADA@@tkޕf{5Wbx=@^J})K3x~JkwI|YowS˷jכ̶]/8 N Rm(of`\r\L>{Jm 0{vRFEbп3 FKX1QRQlrTvb)E,s)Wɀ;$#LcdHM%vz_. o~I|3j dF{ "IΩ?PF~J~ ` 17ׅwڋًM)$Fiqw7Gt7L"u 0V9c  ˹dvYļU[ Z.׿/h QZ*U1|t5wKOؾ{mk b2 ܨ;RJK!b>JR*kl|+"N'C_#a7]d]sJg;;>Yp׫,w`ɚ'd$ecwŻ^~7EpQС3DCS[Yʧ?DDS aw߿)VxX帟AB}nyи0stĈCo.:wAZ{sy:7qsWctx{}n-+ZYsI{/.Ra9XcђQ0FK@aEDO2es ׇN# ZF͹b,*YVi+$<QMGhC}^}?BqG!(8l K3T[<~6]90}(*T7siv'=k 9Q2@vN ( R['>v*;o57sp$3ncx!>t®W>]tF-iܪ%GYbaRvHa}dkD̶*';ک|s_}8yj,('GrgTZ'U鋊TqOſ * /Ijo!՟8`"j}zӲ$k3jS|C7;A)͎V.r?t\WU1ojjr<~Tq> `=tJ!aݡ=h6Yݭw}?lѹ`f_" J9w4ts7NG GGG]ҡgc⌝M b/Ζlpah E ur C&`XR JcwB~R2EL9j7e\(Uё$׿atyХ?*t5z\+`/ErVQUxMҔ&ۈt.3;eg_O ξL1KiYLizpV:C5/=v-}҅"o ']쌕|tϓX8nJ*A*%J[T2pI1Je;s_[,Ҩ38_ь ͰM0ImY/MiVJ5&jNgBt90v߁R:~U jځU~oN9xԞJ|dݤ߯R> kH&Y``:"s ayiBq)u%'4 yܽ yW0 -i̭uJ{KưЖ@+UBj -&JO x@}DS.€>3T0|9ē7$3z^.I< )9qf e%dhy:O40n'c}c1XҸuFiƠIkaIx( +")OtZ l^Z^CQ6tffEmDφǽ{QiOENG{P;sHz"G- >+`قSᔙD'Ad ѭj( ہO r:91v|ɛr|٦/o{C Ӹ!uWȳ)gjw&+uߕt*:͵UMQrN@fYDtEYZb4-UCqK٪L.2teB ˛"ո{Gci`du듎q+;C'16FgVlWaaB)"F,u@30YQg˾_YҊŏ#_f^ TD=VAKNl4Kš4GScѦa0 J ()¾5m'p/\խX\=z,Mw˭x:qu礛WԓL!I? xӤ1(5AKRVF2ɌУլ F "vuhc=JS\kkZAY`R"Hr1]%oR[^oI]${&L8<=#0yaKL: JJl r;t#H+B|ɧJiM cm)>H=l}.^\ݧM<lu Y> XH\z:dHElL(uHR0i#q%]!=t_쾋-, vW~*ſ/,e?IsoSrm_7dPΣ|ͣn/𚃚p9w#z A7yTJ$KOL-aP+;;%+_6'Sr|@2nQ{aK|bjܒ^o(מO80$QxBcXE ء\G=~j{Mܚ: hLT!uP_T{G7C]Ch',ެJG~Jc{xt zܳ'鮱iX%x/QOݸ}S^vv^2M!.xR0I(P 'fΑQ)ۢWP Pe>F=>l |fͨ3|'_iMcĚIdo阊;md^6%rd9#_v2:Y`&US tDkQ;>" ء:9_))wF|;~(XA PLjy*#etĨB$"xㄡʪMc~)j 1駭~բ>XiN .U轋RQ'Vt3,F3,#Y3,kJ3,LhVnKauomˠ_>2h-/ ђ(9Uq EmFjq1jX]DןR24d c̖F4BJ2ᮚ苮p(r%Q 6<$(Ӣ(RvA A-^dX?Xa>EE衢^}p/:F?}bi0>Oh%\x(bdF"F 'u Qx`j#(g6zƯRo(lџŤnE7^k(|(4s\9#.\r= (mO(f=rWmd'rDZ~;o\mkmB`s ~7!GdјCyEߖs|n|zu0VhI|/{}BC6q>HĜ]Xgy G[Ŷ.|37xo=N4wjDH>:&EOΆ<䧊1v@b&툒f!yO){~%gq~.LK78F#E01g.u7^Ew_lv۠M0}qk:Lx%` urJp)>I(>z`{|puB"8#YkrZ .`h(eek[?̱ՒOOc&!dVzMEHH*V"MC Qؽ1Omsz/v0vȌJBIG,CNˆ-L{L #cNqgVR2r뭲⭊ڰ08uirP qNUӛ<|߈$m뫷dùB Z^-_dsz=F8jH˽&DUh+9k̈́W^̤F˖.kL5̻wS"!5<@&] WE\wMc%={_bD&k 5:lb69OBCC*Fn) u{Hk|v;tCl2m s]-$zQpɡr~]Si!ڣZmʢ鉗phw j8\c4>0` R?da,ȍ/ءfQ 2ؐfc}l 2窾ۉ1k;A@z>T+DE 6Хm<쉶K`'#NC5CL]5ݶI5XK.N)Q!>zt?zpPC ¶.vBTcm"Bsp rjﺧK]0/k<'dzM2dk–flE]_vE P / څZg`9r| 5W;`.4&XkĴp 6l0Cз5O[{B-bC\/`m(9A< f`mPіpNЦXn6g5m 7aTcTA,} q:|CBp_uFȆx6ڮܷnZ8dsMS^HэUlq 8\C[n膗:68DkM\7"Ǻzfbx]ۮC=1ÓOv$sY6eX%]Y{⦁# &SlM'iMJ았 t% ~@1c@K?k^rEXws zz.8`hiPܮbC7~n b?`CtjT6l>X+,Qb5ȳp`FMeXÅ0+!86{V5y8 M`_Uw ȗkU]a[.D}"\I5/1o٩|U戻,6t錳"EFk:ZM/!ݛ@pRu Iヵvyne 0=HH3n@.>C@{GP 9::3(6e™nvOσ =?6ͪ)Bppًu_w/m/0}T>CUX\!xl=ZVM\aٟ6h㗶E۶{O#X26.Fٱq1M k'JE%"2.*""]8yܑ4> >X1 smD) ̙TީXfnOFg㧤[Lo)[fLPBRB+x7{{? ףro_nն-2n6 Ym^]IL'M+;U t>x]U5g B(, qA9r;$IN&CM(F+ hGI~Q<웰[, qnriY]3_P${,<\V}7T g6Zapto}PhS/b&X0$Ba{a`W%ATevoYFF"4En.O8ϵq\FOXƀf qbTLhlw?8p@{]oOtsϑ`94t1!F PI;i`ޮMLX7sTGP7^s08p15w q o(uLYQB_dWoc0a#K1P,8]P)\wEZ(VҠQBT^e^0F;)CtT+{`Bh"% !.bBQPnT4ƈRa[F=3}+BVE~8R{3,>0|:,5j358W]>!Q1"6oT[ҟ^T;725Xa+wqlR)<#!9!籈K*:!@NI^S"H=ofLx _lp ꖚӜ3C 4dM @x>ۙZh _uoֺip&1ڙʪ4\RF_04H8@>fXmpLJ5jRS}_D U4x[c) ,`̔Dvckk5Ťã0le۞]o~oW(91ݧ$uxp/Cq6Un9%ZxðvGL qG $ X:w06 E=oWlzN7st˪C:?*|kިfc]| &ب^[%F%LI<0(씖;4A\`TQ.b0NH;ݹ/n -3!: _Jq#Bh^4p|-G7|ڸ=Bx)kre_f |Nm8p5H!jR@Aiߒ߈ۥLFTk"5l9O'ϓl5x|_®&&n]#r̥jOڧK)lsXg\{Md-% >~Ӈ/( [ycy`ðSmn_O;3=Av3LA׊onxlM?~n Θ5 ӂxzPMcVQ@ӤomY42nrQ\'"P؝J7g+#!k{paqTԫ?o?VU}aK q;T0zqaj0"2p؋9~bޏt>$AZLk;3qUlWU Ry==q?{WƱPz;| \;_D[T/BI GH8@"t*"9z%lOONRѦmDVmxюݏX}K6"Qi32\-V_kR(I-wtSJR^m{d a|y,F9$^@mdH֙toN1 < ҷBq/ ۓ,j|z6OSu;BKŨʐPqO K\{jDiy@}b|Z79ߜih(+PKO;!o\戔Sޅ4BFEmBAΛ3>IO j u߿d{=t-n3Pnef9[}=%G*9sX,¬xS&9'E&"/"ncx}"mV5tŘ:wcZ К G)]$mbXE ^ǽ8%>,0FЕ 6vAVKVCjrD25#Lrv?33Iam:xy`|Q'eű^\ơ' .gygSAixپ im41;P^azl5|JE2z=.wcMԧ ax& =`|#HQ*lS<.U׻`>ajϿ '!9MHK:9#s,jV剤C:LIeHJ"M8P,$N;a-zݸJWc :.<sR6 լ$gu4M*B(A ݖΑِ %H;S*ڳJt>$M!^*n3qESfU, Iĭb#UFJPvBgZvn aE5}~2E|=D' ܇q>8[¿yp/9Om/5|k \6xH.Z'OeCD@cq:Y~<1LٖY9# xe8g IKTQ:+Xg:*}.<M{ZH[^>m0G{ ̷hiOO|9Y"mma[sSbb'Rv&{@6; KE.a\}:<]Oyve3h9}E[kMD,5 %sO{킒 8.K?]i/`׎tp NvԻV4|<{H@#*h{Yp/E%dlh\bU:E%h@&SEK [ Ƣ xg{z%ǻViX~鮦w35QE~qp[ʕ@}ZL! Z0!A⼏q)[f &E1K3i+`JG P/EG 4 9LڑKL|`PОnG#|}qOR{Q|2_tH߫%pD?1%(@nfxOrs25rMլf{sk7݇fjӞh2HkeL'Wʿ}Ƞ%>9cSH|cEyQp 'ˢd:,v-us"Iidw>%zM@9IqrGq:&_p3õB!>9'0LL]M[lwWVR9I5YpVgtuZfG{RoZr3ٮr;wW:͋nqCRu1y=㊻Ij z[|W%q0 CJV٨3,ib{eH7 mҝ(3ɏO/̗-=OR\dIoHZ6n`R֑&#.Mv0vԬ]I˟vrK}F9X|FI#g.Gi)%!iK|o}|ֵ7!ېATJKB2Z/"BfB(gdj۸=}'),-iX'|M2roK\e5Pt:*qSH PgƉU'VKξ ,!3`˞t1Rx}fvvPXdQSg6EDT:dׁz^DjXp͇G|X5Q9K$)U?o': .,wؓaՁ_ 3]Q16ZYafuvrq^ѷQT},!H]6{Jw>%wK{)rH+"B4H7-]r}7v8|׾~Us?yWfv3>xpRҧH-EeJ~4YIozi:nq Vq8swHOzf ̙eX-4`TDGq G.tݻgq74ŠqBFf8 9Fk Afq#ϛa$!qNCJ4bnvB @W,v&- 6wCBjxk9ᤉ ,Asy3YޜZ4ΓVYf'h?kNg?҆8oC!IMo:^G10EY↘H:L@D+dˠUHs[hiҕ|֏G/G`' m5p|:9U8PZ7Yݷ/7cs=v{lLHڕo?("lX-Ǐ_whaE 98 (oѢ/Р΅ 7ցl6618ł_1/=fu).s¯?.S[{'g=Ҥ):d8h\y6]t1T7IUV:;.1& ,5΀j:<< +Y?58In'bXIǣO{&V\DŽ0,9f O_"[l:h¢8wݓ19\:f6:+ .3}=uvKc ٹeS<>ij(o'ciS<{1$E[nP b?8E'xv[K+E{,Qƙ1*dcs_Z'407|qBOgYU|U--sG8`u! qGYܷw;ȌCPc_|(RaIBKb+{P.T! =ĦiTob d<>SHr][KqWs7ѝBYǭ~RR"p9dFg|K- obY_vM 4>/]e/dy,8!xŋ5 R<^mYo 3c9(F?h%} -/f;M.~P ؇'k01Ѥ1HIa6Pnz/2ΏL+NhBUx5|T!Fa[|өSfFHİd/D!-Ɩ:;v8`vU~Il2;VI]|Lu>'$X(6 b ?u==O!;*>ohǖVa[|oiya+Cs 6K!x^' 4 l 2JDLЩ LkJu\!`甉܋)`ŰV28!;NfHoQVbapO@B59@mޯtG^n阬iԴ.2O⠪Ri_"6X| w,?VMqܙ7'qpUۚ5Tnj;;64H+/q&wt|RL+T>yub6()ĉJآSz5;)QEAVخ m3 Ō7$j1\XD&?Y\9ȢG:${1`+vour?]8=%Ml%İȖb7A/tpʼn)dk!SV nm=Ɂ-2=|5ʹ zi ' ׹U>8bK0%V\ t!Lku`+=% h&)RO*GUڇvI`b0ο0~oI`b#FOf_$q^!iA9);q`F:Eb/Ec|֨r[RR1҃UE(A.1Xazn2a/"(o9/L1X]~r9_7*vrv |S.Z K9mUxBa"'4T[^2.9ȱ ] QAO=Ƅz`!1Z."ћV'rX-|_W8ʐ߯C{{s10>"=b'K=}|+:)tpՉ8Y4cT~$)*417l;V 6||AFĴ{3Gl& Kq+%*?V6N{I_P:9wZ\"t%>6ES5oà␻,?XAe0bX@ h0[}BUGݓVo !ې: %ƶ(fl>'"Bg< 0H_ AvG =qε^a'NԋsI`Yu.7v$RfSA'2Kcxwg*EQ*e=d|TA6L/_x 'I~|;yj8T9GDT!_lܻ<+ƾtVlnEUdn6UˇjJU;V`JUݵޙEOڴpu{doEչjx&? ՃJ5sB>8 O"ssTVzt&sLq>f@JM-Z ? lf@v@N. ̞r2 g ?K@BS %fee})\.?}f55[c&-Wgay_*ɨ7w6I}qww}U~7뭱4RߋfnN뿛*Hmc 6~\h]e+]=ܚoVbmy!7R7wPuqpqo{ߟ+[w_u3ܸ'AqC_oB㖟E-? k[~;vmcoW]M鶯4<\h h[$k 1"j {]]O,dE~ZHOge0Ώe|Bf"m\ Ru2 \ )@+q(7}* h% !< 5IDK5AhDK֠]4xMvXZhX1xҖ51Y njsa*&0Z|ђ8 CGV[4xI(#~ t>lf@ Vi`D~ڇAځQi5ʬLHuoma_u` !>Z;I"9YPۊ}xb6Ia{ sckG,:\n%Ld="K-us0d/N#X.?_םh 2r-'OID3d ֺ"?yUAmE}|Gr@!&H2D+vVyKZt<c$k#xu;ʍ-`Zi3)|x!6!%<@fpѻK2Q31pxFP>TU?!$VQp7% ' c*K "U8V15> =҆#xɮDپ U`۸ہt=|X!~A:"W vzZ U{ TĩG /z!~\<}LY !!E%ҏ:H =VՑLf*n6tGd#fR*c ^,—R9w=3}resLV̼d"I va,Jm_u)d靕َ&4pj2褴ƙb<;dM-}T\%.}Ta@ iS*XK!\+ xJpΕ`p~mΖ/K 8Wu១btM4cxb(9A `aRs ^d6GXA1?V_mW puȇ;s`uih F2% [U۴"qkjXS~vWP~ RTMrQίE|S`poSOfb|RK> 1Gȩ f,( =X?D5Ir񾅶X/SR?|a|>lX9t>sgv"4z,rKrNS)`Őr\+!!:ViI[/ѩ آSNE(ȊD1ca$j1M (C>ϭh'{ߤ%E8?P%ٱm4[$++P1;ڇ6UNE@Uo/'}pl0\neή]kY0Cgcu6/!_Ɩ} ' Ўa 2+ဝ+A[QA%ع „&䒿e;]0:|7IIc(s' ѫQow#z)穇`8[KBc>q,7}VE-Q8W70upAMytdx)6h?āUEJI-tq5ӂ~ яr@/1NEUyzǖQqrGm𻾋-]fT'.2Lυi l-傊zA"daf]>HUd6O뒃^s}v_+CX|,l_B'9Dcuu6cE(T#C0Wg󵸊y!1U:pUC4@"{u%[^3{GGL']9~yL :!czbv}2.R`Hz)OfJl˪VWF fK6%OE-km`l[<-})xǺǍ-- W[`wdй d:[mS%MTƪ?>={2]I)|Өq9| 766fjek3X$$k4sR /:ǀl}9"A? m~R:aV޿TQo,?dڛ$SJn|7.+.{pWEqws]]|/ǫRnE/ RۘF,nC_+}U}/Ǜ+3߬VͭEuY]^VWusxnoU!ryw!׷տ·sH̋P4o`[Qj0ւcz/AР]I,Ѥ1HtT=cKGX iצ) e5 'Cj45<Ui/L6I+Ȧ!6' ҫ{.9Uʍ0\jmvL X1|wrO_KR3V:<}xXh//T3޲jЛӼ8Ģ8GIBxV#NV1;HC78NE@UMc8>`Tvwm#M_Bg7N&ƉwNh׼Co5IɒULLîuvwX2>BUX(w ETFY:\D0~k#PmRCUl &1-JE ~dI<y^[w׾XoO<&t:1H~}y *cDzB Y~V@]x-2}wP)e278 kMF8cubIyVkyF}&Eaܴ ƽZEiJ2 ʨUD%_,>x[f@aI0ETTzZj-/TQ'FhSJX躮 w-÷/M˲ p=pcs_00<4Y,NS? 캒AcǓZm~1Px.8 X<7s0a%/6y%nE4-Bٞb6iښ5Y<B#.r`Ug,*F0oYX}( d (!uUD{oyfjjZkMvO.($GA/(؆G Džsb<:o~~ _nN)r mS-g-wgOv.ޞ< kYq_W<[U2\KY"ʜ"׮Jwc\Ӿso?e&FFՒqTVbM-٢٪]4DuIڈKTŋU wSyѳC8"}ɣKmEbmUv>N8(b BBETLSLz~kb_e}!:=|R>`nʺWٯߚeV~]J3iԽGp-r<|UGR~cbSG ˓8D}&mx-eoa}M[mj, bP%o,<=Gׇã?k`VJ .öP K^6] nw]z@b"ּ85b]uBH簟"u"p#+bpmُ8nW79tHھ/0 fQh>sihb8 Zn,`qx(GdK( G@bm,)YݹQs: 2濰0vGL^kʷP 4Z{v[f@3 PLBQn8@1k5@qݗݑ(wb~DЗp$||Dъk͸򢥐X8R12q\n;BM _Z-0:ֲ FL@18}bXh5N)J vd—s, wK/l ^>Ωw/SBZv4 4d]nകH&3zs|ݸ;3dW̶;ʙ;SVP |[Q gAA$pt"#yy'G "[?&mdžsnw%e2 {M _>skM[NU¹Af{mPm Oۉ9Gp/8i[Z ]ݥ/ndjѵ.D"7y[K1D'렞dwR5PlW 6[& 8"e!$H1}P;a0\P6Mw!"`Z1LeY#2Hv3ӗICxpT T>VوTfwFwD8^wu#G4 |pdqq%2 lplu+<`j6Ӱ"rlX.J4Ԭȁ)yfu#KadDĝ v|,K2* ֺf%do`zٟeqCmߥV-`0)a5t^ME*`cu%xK5G`iۀR )4w j,Q)S]&1*N/eÐcvd Nb2n!aXi\@@فU7h͓I97ZG8zQSZz`vS{1E̐w2CSYDx:-(VթUtd]* Xd,eӑ86ި =7l岬_U_}"a̓>}nOd|QgciF{96)OQ`LFJJ`#mР&s62-RK+H*|0P,G}3₥ VugIj4_\%O\o $'?tH}/^½x) <aeA nQ{ {4qG,P?~/t_/$ͨCKlM0@5ru5FsEB:_"D_'{%~/ʈI>OΡRqfOf "#iX \OY]ד͝8fn Cbou:OrVT ]Ht?͝ǫ4liJ[m $j˹W]_̇;УdӴ(O1yaTϮ+X^'OQhx#ֻ)y@Ek٤(;gd۞0x%x\nn%|yd:*Lgc8%oΊ[TW^qq*+~a \t{PpyWi~8 (<㌘; зO\A"0nʽ z R9.M95ʺ,HzC'L{hҙp 4Ag3Фpc[KC:̆~ a{2E`0=X~+u 5TzCtM; O=tmwOe/axK6*o H^k G~IDkiwX Eӧو+-q]iS<ь6=y>rg}9gB6pyD_Qk!˱!"f1s|'=_$Yk4-Xa9$ l ^v0M b`JdgFQDDȚ/"RОM0r݈lYg,,6ZјGGY׉`I%Q=x էƧG d_d4 $ qQy5[x;`cTL{phMf+8r7 UqSfo&Sgpv6ʺj%hȇk8 XƖw6=5 1|dAˠm\#ꦢ.l)U*i6UArnzІNIByVV e[ʶ'G(ۑP mӧlNe{*۵[Vc]v߾ݹs]J-߂X=|?R-W ޞP{?B Y%قPg{Buv$قPwPw B u#ݑPw BUB-'ۏPoGB-W ߞP?B 4X%4؂`{B v$4؂pp B #4ܑpKıqgrK L)7PL5z|gF]^9()]崸h"zpɫD]>m>.s\h=FF$`TXfeE2#hj\V]FF=$*UV| m>gN]l^:4+Wus9yʼtq'2/Â~a3 =Jq 줹rK]8cbҫDxvca`hEO`V UDت yhK4q'NXe)~y; ZEkI"[ũ飺urT,?W!BT.f GhE !,a [DZMW 9ߊ,iT#ARn _$$=JpRPvZ[)w+aj# ,*)ЖrSJRmlâ9ze@Z΁yrgx1>VMa0X%-ݭ+n}:z=ea?(/=`pϚ(jV4]NlvM2Fܲ]zýP4?W6\2bPͨ7R*oYU}px{ 2t[R 3hm'\ Ny|":\hJߔ1 A[e^ r 31 W@l+ f6!HJIWM1L"BUVN2 ~fBbgsK`1)|"XY@z'!=ZTMLKEz$k@\5k>d%YX5X||= \]YqHPa,w#*eI>b& ӁuÆI'SkfWq{TQ:<(B;\k 89l4fj&~=pCS' ̨W" LTa/<Ye@ VxWv֙%SRa֨Mb0pA}.Br07 5+xP-`<~SP/'{@>N/2 `+H ǧ>aóߍqtR]B^{#Cm<0M_15?Oߒ/!at{nWꕇF=T?'g .t-Ƿ >DGt+0" __./!T>TN3'1@$^_mTdYvjD7":Sݶ;vۢ'̀N{p}3CHZjcH٬0C{E5$ݺƐ~4SeIirD<WI[dOZNo2`<3T6(1ҦʓxfZe,5chx0;;]:2Bo_J#f&N\2l [D7g!lP^(ZIeXM2gb)((#"nɦ겈aΪF).J&P}M #k<K_G݌Ctc|=,jfQV޾ ^% \(u=sU6?.Ziy%v9jőBϕ d:Z%%xsA\6 xm*x]Q(TRZ%)l1Fk:?ocV|YM p6Y&9#WDb#*YƑvɬ89:0ꍵ:^Ӿ62Zؘ ,Id"Ld#C<.,'"Ԇ)EsϪ˕ e%EQz'E&̆{hv - ѳ&h&y| w~@,L}iLݒTdd0z\%tK@=unf$ģ9KlKN=+t DŽN$^Hc+aA8KG& Vm#/(kmH%U~?, SfDQ IK[}I7];{ '.G]Stg&2jzOro"ڏ~,`}[Uˏ4u㯝Ǜm 7 ;S(~x~&+'P_6<_7mM|S3+w]Ȱh+&սF&@>i0s4ٲ߱IC~,qP=>`+^gMCۅR#f^9(?7˵Bh36UpMG5K|?%VaPT>I3͜g 7:ՔL\-3 mV9[hV&o&i|"VJ6_5O]:(.PFIhpeRh[x?P,IAM???vњ8:*3%Cp.eF7RHtc6?/]pb(Ѥ,QJ僀ύJVK0}5zL7~➺\l>dbc(դ !T',"<3pu`V06 YV_4K6:chwbN?muQ2D raMCFCfcTc(.pTY} / &9Ƒ'*1<5xRB D|PsJY$X8zaUQ( D$HqE!31$XjCjOQ-hJL0`ǻ|k(ƞG/t 5="t SkՒTR@S;IaSnPlpP. JDY gU@@bvۧ~鱷E0ikqs4smHt>jË%3Kpl%tJ``,8N 9քS`Qtkc ч6J9TZu,]S큣|QIGR'eS꘍9+$^{|d0"a.B%H&lc~AМ$4e[]ҥ!F2 +|x"P"b<-o@1%~zxM[UR3o0%+xvIZd):1Z=xJ53%aҶNS6W!5Z Xt!{M]pTrQ &uWN~]b-_77ҾMry(s֘Θg) Q0ZJ̼u1i>@ Al6YMp(fFv2òZ@nc[n7^bAЩtQεʻ1;<`5f4+#,ǷgF^+ X6(kXI7Z>i!tfy3⬽UIcl_>P,AH]?QX rlg9 <JyNv 7/u!HMy_ XU)=U)ebӋpV@xATS\ۆ6CZbø-kN,](K ]آ2HJz4n;fQ[e[xs,%q)9dFŽ`~Vhqzlc%J ZNFLsB5/%X꘹98R/Q vD/ `,=+!B-<\҅ؼn ͙̞[Pஜ0J ʶô,\1Ή5KJ5>spδP r oLUiSoxŽт8Dt~]ptcJ)bV+%Yo1K ^2;=p9JC(S\dME0h0͕>pX Ba٧|vݡ jl)yvJ?h6Z(^XmǺu4s=K,g(w4r ;?E@R?`7+&2 90<5 ̹f*&h+ LwN;.B_9,MJ)b{,䲎9$ؤa`$բ՜Nyip"~E:~ |j{ J R0_hSHGb[vzDq{.Jp/sP}ڤrKQ}ɪ ,9;*r阠oνθ]cG帜~vY3~Lo>_6֕ /_ ATCJᡘfjpSypazr]p:R”Oj)LD>&ҥrwGKUFhY&HY1PRf)0Ć'?mM,1y;K4j+_QDu{-׮;i!SV~l׵b8$ %=~2:z%TR/Kbf')A*0 1&g#~oE.%OaZW{!x%tm[O„#,].Dc~QvN ~ؼ g gZ|d):=ErQi2 {s`\*D+ޞ0B&NC]\])X{σ(LAQ|̏NRݾ4ܢ,` ~!Sbu+Z!H` υҹyicW8D#Msc҆U\ L_1MHK1x8{qU<&1\ 1XB{&!s 7( !|v̀b0Z)nr0Z=Ru]j0*yM4X9 `DhXT8bR"<Dž1g1  x fYSNNow"E]=as'VK<.& _iU'ckiK605wml8#;n(%~PRWk8E31zk<~GDRҷ̭sܙ/5i+!hM_l,)Cb&HzGz`D0֐YeX4>c2V)B1ъM8x+osӘN2`H$X\d6KF[(JEsv _Ti6%E?{ [VoZCghzS"x ޔ#8m6X"O*@<kԱxO59*d T.SfkǡT8ԑUxBgfRʖ<LOAe1 \ZL9QeE).Iu+^2 Gj@Ly<7qv|@/ W. eMk7k}.~iڄmZ?~ZAW_8c{N :y+7!*ڛ1bbƒ/'G1փZӁC ?}ˆ$ -0.J-b=?l濾8|t"ZE|h_yȹjΧ6ʔpcIy!qy}p nY}D#Uܥ[T|I'睪]hG~*ߙڿ}/~Y/]F}u׳8Xu;(ȿ/ykugk=)]%rRux60շ^Ci`TRǃ e blʙ;E9'NpL1V_8xV4߮N2G-YcbR:}K o]pZ_ g/7oxoOcyS93 ꠸߫p0Kk[2a1b%$US|s=#2 b"G4@w_wi^?gogƙ0 b:vQxR?wVFkx;TM~}?u+H1~mc~Fm4{JJ!仍a6C|xg *5bLwpZ<ܳ߸9cGY=hB\Pk?2yQYߨCu3YB:ڷI).8J-HZKհGim ?שiS/j}&Ns?)ou<ξ{yE _#A ^|Qi+탐NRf aιG,:cL_.8V~v&$RzW2O<%&K1F@Z 0X9BP, ZEE3%y{>Ph\'}6:{|Qwh=/ f\e\njOc8zZVǧF[U`dEp£ JNǢ = 8[+J#@si7: W7W><*NÄo2>5:.'<N򨒯%+&ES|nsxD ׍7GqvQR'vօЧSI;eһcϭ]zlޮݮCzůՏO~0e )%C aZ{9_iYl P ͢7;BMaPu]c̄]忿۸m vӾ5epSTթ*c|VI\ _(^Ig^Ry)y ۍkU|+Jhη9U΋ܞʑ19R S+PHBˆ75-;٬f"te0J=[6˫ ^JoI'Ͻ&#,粮*xke~$~jxiߋis%-߀KbF#co?"K͔UQ$[N/ Anz'uwGN:/0O2}o&3 ?͆_ od.88?mb-GEYcF~yҁ"Yv3{{rЙrDžd16r U5W6Xy/eG5s,o,r&^u#G[8e !4Ǡ?~s~5H3nhdjчf˪NDK6I<)Fh=Nn3] ܜnWG535sk=5  쿦4I1=K )[?x;z|3Jng?ɽ}\xCWtG7

'Wh]x3>x]3'?흛nx|tEhC\lc "uG%6ultr+!_ u$ٻCȡ{Rz<#Gg4WDϽ:.9 Kj`lYJ_nw^ќL*z9h+{Ջ8zhcg'94NUO*W>DU*x?6 Z#IrUbeh!”T;PeQ%o五qhǃAg\m~e?_Fv2ijj{>g,' :fV:E\ 3qf޸ę)]?/FfFǦ{U}0nItiY"[W+3$D QzrgƜL.GѼrطH4ԳD㊸+tTgA?Qv&9- OZ2̦bhc4IכOp46uT";NSa%g;c}qJ=(8\xtѰY>e tO: Y/JE]EW{ gnDPIo1 ?>'PyfQF$4rIg:OgwųibRu:j#TX0уbeW~~k]Vz,p7:kXeQOR ;9iUִ k,qfJcqj=KTXg!g`Z4lf6k3vٺRfk`K^ "?QOg^uA`ཌྷ szs)ڋaMwb[0_gVp8S^hoSX}0-r$VaY a{Y^D8$! i_`jD$".' J,Yh[k8Ռ7g38fŽ]]?Z#==*U[ʾ5݃4뭑_hZYψ[1sJ۔ V)smYƌۃh x|yO?kh\`18ȼq@_O+iP+.fPRÕ10jߏ)??X}I`5{[mK!ߧУSAT5!joWrd"#kpm{5޺?6߱wA)| h=c|}R\XUu@C/M|ۉaxdH߁ONaqPIxr^2 貱*67T1 e@.N[ȦPDgWLp<*5mx}^?Esz*h6]Ay&^Q/CD@K?{qnC X1tWWˁWJ3gmG+͂ziuLb6>2`./g<Boݳ+.røإZb-2da 0g@JQL*E-;^g.Mez [~ɫ7F0Xu Wr Jj̳dG(8)g0-qb7\KY i=TK@?i JQ1 C*kz8&Z@W7b!FgשZuʟ;Oz#nҥ(0d%J+/I@FC/3 DLԥ"c):C|%S+&QfX[Ni(o1:!3\&j(' PF8GȧN8H$(Ƚ+$WLT+kzuꩳrE# vP'1`P )j`բQS4X֌1C{>jҚ &F (й*YԷ/z㞩J&#*cRbkS+16p ,18s3AM^z0.2f3״0(mb>0X\M|1[t<| fH|yad<*,*-NSlbM'ga}H؄HT= TceE|8q?YZrvnwK;f.e8JX OFGJh] g˝SL[3ҍPRwo=obXv}khK31yNoy1[;޹cz2eT8R>@T;( ui̔hM1[`y1bM" p}m6w>g.m6i!Bv6aD'9N+jޓ_!Iߓ7CWTR:Rg xe!{ÑS60"Z0lS!ֈr=#kmNbW^u~(\`v,KwPZL8We-:Mak`ᱷ"Eb FZ}xg6e5猧T0ߤR )0 1r cH1z#a7RD(*5!xg Ab f -*@, \X3 0D)''qt+`Xp̡ݠ>xЧQ.N{G#wZˑũ8(*LڨZ]TRJp^w 6O)v&HIA"@GD9PCl7|JsXrRE4 d0Q/zϽ&Yʃ!|>!\ # fZ# 81(4i?gZx6_!Eĺ@nm(M1l%$5))yHEQߜ̜9g|Hal vwa P`JɓIAa +nb{)@Lp'vՄQv2͢rT5Ҵ *:% XuH7bAњ@9MEe/ھm!,lKA6]Z_hi ]΅d U>6 ܙ,5Y$Fd-҄Q !b IApIx "wl2; #ّM҄N[ 1p]zw:.9سq;'-En fGJ*$ !!,wX3eضxx>\}(B,mz6M?hxe`HB XXKQgB61Ნ>zUB,Ț%hI_N0on0Ln.)&[{np\=;gGCh#q  q1]M6U }wS;PLįF6AY \` T`J8nmn'}sAS%7u+#]afw3հ{.0 !:a5 fQq9 rz0;#p ED{տU~ Nw솙q3:w[zeV8!!dT<΋T3G=nNWj?*XF\e;rYwvXm4uŋQ(-m۴96"<%:XxpLGү5Kq: L[? ֱ_^ZmS4ZE#C{M~b}&O&(kէ׋-zUOQbrE,iD⒢K ww[ # k7x`> [@XI>0!|| ˜6zaXy9YI>6r~_m Ax{Lƾ5+<:h :*: EhW Fw WU˺dW] KMe|cyf;Eg&#,=ab"M/``AMc௏ !+6z[j;1&]+||T <}c,;gQ$;IPW9i 9.CB{h2,A h- bD% .' j*~FMJl*lf6sJ Q*/yRPZW FE l*rb<`!ȫ08jf1Ϋa7T/Z (;^Ssh2NŚ:D0}{I+^%^!!^Gy8?/ '3Ž+82i&H%r`( _+.෷?ɷhח)L;>ذ XD4b%e Aa߽zQZpzx>/j?zV8|?%̉R~q<;ϦwX5#G>f\Y!t:Q (#$ZA{EӣoxiO_F;x1̓W!Z+[Fa>N ߀Pe~+1ͯ6YN㵙Ycڋ)i;΍|v?{WvEeQpǓeeX4lǾ:kz_ uaӚRyVvv=o~ 6$Z.#ȼҝwQ<'1uFnGlhϳ69W;;ID+ora>9sw$ 9x>>|r8Lr˯JN@WˤWXV3Qf_8H5`{2)VrJqdo;TVsrn {R˵#̱| ;[6wM'Bj@Rs[m ޖI|ZĪs~e #S/Wkvl>^UN/G[  qhp9ZC[C۞}~+:hOfg! Us_sUۛv{hBqgmvu,7)鐌L΅CjnpV`tva3E19gV]۬֘J>X,0`fwAT%qA+D9vIq.S$Fј3.46D2g9Bq"hd%&rNcScY'oػt˰ߦ~h5&TWOZ\0]I79A.{[A+$ycy܉OwkNdㄚJOm`7GHN\uyA-:Zc QjE~T~kqS:pK0e0X~H r%|x-?ZcQ, 1z0˲e`3&C(凮)ytG~~gxvn=" "waP?peYFj\Z@Әe qLNn)}˅8ZNч}XԢ1>}E,`tm]T%:d0e3½oi[,=\J^:ߛz)\DL#! NgO dKhS[[7_^*mmI?טrgY kv R!_7\v6`ߝ ϝkbPwʮL'ԊuUʍk]{uAxK}j>B*f A垺<:``|7NpK_悖J -j9qq}Q$D" ol+l:u#v%JjpY' hIHH!%CpCӘ*B1*Q1kj{mYۼ=0I/&̯)twFң)tDU+ FEc{'Q$ k[o? O XOh~Zy.kqV(˾Kxk)PX.I)ҭ}L0Di_`tk+Rjh7'w1L=#i@m{4wNC*6mj]kGl9H_zGj`EU~cr95zfz|!dz3Jd>9c7 fdž l):ALk|xmfXb {e^ch5~ț,Wss$ pn5Wi$ӏ=?CC (+tGF[鲦X>u2S,;)CN֞S7e -S{LrspU=E><>mSJOwÛkP5wSFKjg\:tѵu%ՄߖkN}1'/6n3xg-5aϰ⍴v\ dKm:Êk<~Q[-?zW~CӔG'X2 n~u -l;xGRxk|U2Ȁۛ^)MES`n,;JusdG%SM8FnS*A)Ă# \_낖PYϠ\` m s%h]ߨ|leKd1> E'nRIY B&NEʑki6B"p# M HbmNn9U,̼e6b7n8!|p~jg =?R1_o?B|Bcy?vN53l1afsqLycD8݈^0hHdG쾩fdxyWdD*D&qgN`^wP ?Ma,5'2knpy k.h?P[OPδ$_s}'QMr 9 `F1*Q2I8K"+N #H'&6FX%K<.diw\2.s#To ֤ƒTMV7.DsBd\,`MR[LƒvkPZ=׸1/k\G"gC@B1{.=KBJ_:VTynp/Haw{G%?uunqʊX"HJ=5tyS[GNm>ja} m$޹ڦwHR6m؝5'ƕ3{\{wꋢB=Gh36 ( L7xp0X/W4glȑ3W9H!?H(|Xxۄ׾R#rD.IiV7f(V&&f$}f˯alߟ)6y:߮/^7>-&ujBl)\c@lZ}Z6,W+hZV>}_6mgV ^^^|X{H\"|0p&v5ːRywi<j}ڜTo>m ~m77眭g}^mSu>oR[f۳!Wqȇ C z}4vOֽӧvC*a[ĞyjF{Pfdͺ"kƄ:MNzES3OvԌޕd!p>؂]zldy#qg GOx>xu>7u1[{sh:azsh;]Ʒ漳J4"e+Nsu)k_^l};>v2%O@H;£j6NibX9<OZ_?35]>U=>o|yߑUK7Vӿ_=}\etH;W,vdfCêvS?>=F}OhK;NjYT*ad V~,> w^nOv$CbA䀼W4" HRcYIrD*ȭ#uFJ⸂*m}p 8B>x-#F¸z(#drΥiq"Jq+jY~<^G/%loMu%EM>rc'mOfX``)2i%҂dc%_x 5U%$% V!;*x/U<|.+U{R-$2ŸޘL?2)eeJB%i?^aQQg-P]/ @ Tjɢ/CtD(Ac2e`2{(XKM.}&cjhٳ6IǓA7 8;ZEX+g0<|Pw8jՃc ـ!nů>PAK&G xj-"Iҟ -9_٣oRT\z]W!b90$Lɇ#Lpct#/!w8o3}[1+q䰖ƇԪr>5{hZ0}uBxPeتa3?W Ր"B Ħd1E23Ȟb4m|KMe &A!EG:^R\VZ%i6쩤Rn4NIt|@NZ 2eݍҲ=^IAU,s#Swen-eh5aIӢA򑾵ZFi659՝Q`  )R<@ؔyք Dma, !,˒i_T0(@d d:H.,d%u)l6pY/u 'nm5qv}dZ#MgpuW j5ve>QHRCMm޶۽O2 TlH>rA{'rRՆo.+JFo 8%Vm408&IY$(R_<^ _ŗ]/ Z DdR8!H6|X*|qh\wex~tqfkݒ?6KI cI=O;`Mݾim$.^``HpEdx$ި})t,ڗy,n_a  H %RZ"Mȇ6^˲e1WT?k6ɀ-I$@-NRPU)0z-1D.n6%aDR|`d%$y6רY[[6UEḵ ޳KT|M.5ic 1 g v)$ֱ&YB!$djIA !@1GϘ#=L]1G˘ĹKΆ " H;],C7])Cx:OW<uko$8Lm"UHEiJL |,"2g 4}HĘtlDHT:o RIyRj AFcIaQ+)k0sydPG,$b*H>ޣ;\rg{㋫x]W1w wqh|܃w[j$hLh쀆0WRCzf>ˎI<QIA5$h m7ƗU ͝]X̽∶G B-DK1e%~Ĺp38A_ :'.Lɇ[ǹh^mn#6NPu)$!M2^Q22s{fo*8X[ư#2) T4JŇC!/ӇQrf0g%2 AA%,l"ijX B!M5?Ke2 CʀH>)eSߑ2QeMlIIf/W$a && @˫ƛh=6ߙ] (X&y{.&"!H>pbJmeؾ2@1PxP!RN$Fbe2xns^]TP&);^h%j2H>+oWa[Էᛸ9m.@g&VeP فs:>%F>bZNx1b^ _ jL!8oSJ$P(Q|=ǮU KjEPLD|9KhW%gIY<~ Q`73o['GNE[6DOq"G4xwHO\.宵]\k{"o6^j-j&#,@FR|"hґmmnY2+,!d n왉{>A AdBfP*[<"4ǿ,oO .)B-X8b!9~L#,'aȂ&4TO P=% PeMe[}7yJ$ 07HENsHX9-0zX "{{BϛbG{&`H $L,cs` ŮeX <([1H'GbY|]$K҉ x4j0ru<;I'G^Xj-WHaiKs+rHq<5`U(`Q;!i쭲^ΩHݑ<+Z@a Ndl-W*KA.\Bg&% }okQ1 NBI ] :xuIw :yb D)EA7Gpխ(o6&Ae2H>PV_*cPIߖrȼU!.`Ph"e2qx̭k犔Ǥčt*Lj<$) v~j{%[3~w =jNLVYPrwKw!LkD|R y`|#վt+H˯'?k ԆK+?Lгݙ42$u`񰅖%wAyd(]*J]b}||{+=t?vt4O_Y5)sգd-KLAg psfá5cϹa r*EWȲ"?ͨЛh (I~jybw@!T/{# #NQWd";i-'#;p`5~o~6;͟R}͋&`6 cr 4;q_ 4O~V @dP{bkU F9fGF?U/gj~=ЭWy.QX%lm1u5[e#rl}"RQ`7Rr ~^5KM1 I{SuXuwP2^ EDg0Fy\7'긲+OW*l SCJWeqs'rp'g-$;Ts4^` =J8~DJZ)>~>~{/DF AKA°')`{nfӍt:k$)c9<,Rx2uz+KO%'$DL"1N؋JO~qiܒYd5>4 4R1.7=?)a\X7ƝR pP:7FHIvjZ^e0 xMG2tBx"W,0-s;ky9) kA^4ub*z֧gm"]4fIk5'U>qQl"q=[g7[h[N'fHvahdjㄑJ1(8B[ )&8q~\8B !f+st-EWXK\dhYO&Hf>Qe5D֛fӲ乭|~,˘AwX1F>3B'|0v/|m;qɇ`Q||{(E!XT#caαs:Iؑ6""+,eH uo*ISz=N[peL6F 1ՅOE7# *~m|BĎLrK7&Y &ֿJpmCg5ju&k͜WfÌN1`+%KQbDE?&0Fqmomd׃v oK>ĢXld 7@]mKIRKBzZ_h0WnLEy8sV+Hs_\5,}6 YH(* pd*#p#۞p@4cr\xˑdG)~t=^鑅^Ek^W«NDc!(1RtNcڅ)5kF*: ^EV)Znl8+%Tr&Tmd sJgI8r1nNV8BdPd)ƫ`.iz]ٙZx=`}k3iKr`07qNLUndyozor,/T, S|է|(Vp,'=7 x.g|~ЛB@׃޴!vIټ<Myh3ߙO6[ ¡q;Z2ûВœ6,.-_wܟdZu.ȷ\pl>"YpiTڥ-lUg$2Cm0k)~QKRx7Znoq竴Cc%L*MLVΤݜxZgR,nL*6ϣFit,qNq1ޚC-H@yGI[a\@ӯu;Jg3/.xP{9VI&j5:3VKWBЧ|!ᅔWLpIBѳV%3V5BʣXpZVD'oZ32XgcWl<ղ0di˄vp6u|x Q(0%ZE) XqZ9 [3P={U@~5+!e0^"vl,-5F]ECsZ]5Sἥ0}>P,MmMb翇Ё'3rBlDK|WXo}O۶Ӯ++yץ';}WJJ"+\> ^J&Ҙ(cJfQޢ诬EӒ9^I@L1bQvwň#V{jh;Zt6"Kע+4" L1OJd )|\zΗ1CDPV[%-HMu0}Z%J.RHOg\Ob( +9g<'_.-:VbD]& b+L= `NyrzAlE U؈ MVr^˲s,sTwXɹZLH΋Ky~8DCu>CFkͮul0?{w17ԻI)_ehBkoK;ҌÉSP![P,zQcr& ͽH&^c<4Q/$ۂL…Ect3Ƀ7o6y>K˱(&KdAPqSqC7*Qu1dbu>@Nޮ.NCQl;[ N\;AYx(EQ8}?}F\b|uIx "X)9s sG/SBoQxFr̚Ҕ3L@gt+?``E5ҭP,DJ+9d+&E"Rop84!"sgf#@pӌݼiͰ w7joKk!fW6K9m҂h" .3Nr~ә*UT&h>y< Nhƽ\a;wflV9C1.IthpA:N? ? L?ƫX9"\ytVre)@+ύ&h`*Ҩ/̹}iz$|e7~0'yh {{=o&;4+VlLE=r5ԩ"\ΞG֤@%S;!ʣ~ɞ75J&0"DRXXVx2Y{Ixa9GU-gzU`JL^\*Um8%^DpaGKOO?)Hgğ{_px. hIO()Lw*&1 KI_(E|LF98t8 ^KM D~Wx&F_3L^fCg:),,N>˦HW('D<|gC< }L[M1Ciep1digp}ߛ|6t̛ 5 =3!ʼIũH?PsZSza:  y.Tpo1m?=U">E:Q7;;? f5s"M(нLӂ:յض6mᷖkd8ں.6u]s㈝sQ{~~\^R:{Bs _*5+7#x友REV,|W v{7?atno Mr69/(:MM!K2wv[NJYQR*vtś:fͤUl2ͻsSjm,e4'FޚUTz>SUO5HT6YN-E +(.egXT C6̢yrGg΀щ{ -rd+;Tt)έ2p1̑}2hi135$u7ΗХ]"N@cn̒\2f !vh4efR?nKTg!9,hIal8P&@;1gYA`eJeCGUqj\J4L hA|ZjZD,Chkٞ}kg_n2Rqi)4X%@3`I.T1mTDJˤBVeD> LgArWb v(PPH.hUaf|h5kqf]k(5ʀ[O5!`@6H͏ K Gh22J8  L-*4 [!#( 4ИGgN[r̤-B¸`. J#a]\m &H܅FZ4Vq&4<&X&xxcDQCD=R ةvG<:B#* "QfX(tL(JxPyF Gg@ZlX0 bąXx(A@_;Hp eV1/<2`H ,VKpJYйeHqm( &4*t3AŌZE `&b6q(yLX;u%c1C2Ff B((ɰ>&X +#0XzffZ{H_!,"^䃱 $|YL0g$Xv2MeDL--t{SLS᧜2EP桱qVQLFJA -50!ЮIb$AmU{5AɸC@zK)PPHY`єihFf5Y&ysS 5%gM]Yyta0w*dS9pB`ƛb]JA` e w[*MZlYBPM2ÙH0J8 LseԸQP5gv:F8$(Eydo(_u* qjSueJs.Q))Hbm LcSNU 5DDiy"J]ZـH ްnweX]6CB |[4qpl(<_hG1!.E5P48+ƈMNdvNRB!'̿@1`,bg =3YuYAӶdz|GAw!4%MY 3 ^#xHu`Ky@8C,}Yd:@rraU@30-EMp`A|N k A-)J#H V *ʘeNÕ.5hyT ×Te2jnuG͠ ۀ8 UvdA*X?<{UbD5! l+d@l$S`U1~v?=h󨓹HΧ,}'XR oFP"Gށ]JSsځ^T@@P @r! `"rjI7j5|2 lN%\PD;`X:%x+ lw a3RD \j!hk4^$7e#Kqصj:;M 3%)AvD!o, 4\-J>C0l" DŽpU"ΒMT^9C&ؚKd) MGJCӑpV4Q8FDj2ˑliTjJ1M?z+"t j Nn%,@UC}o^KCTj!.mMѦqA[;5nn yurX,|OǴyιI$ T42:9 quFJG+48z:6T`}&|v9,:Κ*5fZR!ΣDjсaK6(O#F~=~5( }˝ҌТ EzI z XmJ Ce0Ck9AD;I*c%'R ;( ,4**MHz Dze%V=fBV( C XQWTC(M@rtd13 ahaob'J, ItMsƫul?@ )";i*C Y6R FxpaaFZ6@ZP8EzȕQs!1႟7lqz@o˰+Ɩʥ XNz.ayIj.`$%F`oF~=RTm Fwwn_ER>=1@?A-~YJf2JkZ3 X9Li&`A,s٠sv`Du8ОtřUFل6"1sl&`(C4\F3ll0~V C.ƒ Xemjf fT)g֑Rj. l`z"X}$HfVI3J9xLLtsYgj&`6XIv DHXe XiվtjLjk63f=?UBU ּ^ɵ[|ˋ/. +<ç?]|j@t_grE_#/W5,۷[\F6XpbjNJ-fV9%\jSti/L3eAKf9+|+y:XL0Y;Enyv .,"V{)%RY.N[7;kl CzQYI $ #찮 ~1lEس.O͎.^#ߕ}= i:}{:=ӋOkOgeM2F|g˯U5Op [Ň=olR 5si&;.䊎Aƛ -Pj/Oh^ܹ$06llcX}>m~:*\yiq;Azq>^|cI_Z{LpvMMȕM><C~Kik&)\Sjt@Ex)~K"Ə|џW\Tl G~) oeM+=b0S*igx|_`:XBRv&`W).Cs) !}/#L i`UPJ% 4^NUM~&`bԞt6So?Qa{/^OjӇօ?Bb. +~O?wUV[;b`u@ʙ%!$%f:dy9$>= /viؓCg !sJ!#nѯCyK0Qd[n߁e_C?N Cjp8޻mE|M+mMmIN\}|wȹ7O.:_mFih>?s>g\]w:ou;pG۪ek~s/HDuvn8]`!>l8C2@/]"/#߮&|*hcboI]{S#_wzO턻>xc̔u^lf"Զ_&n7 ia`^ ܠ||`V <|fggCY)W@_p'g89r7%BsZoUZK$n#(m ;tk+f}6sFWxh=MEe%[rmH!VymMdBDM\P)&#}բ$j@m%.TT.ZqǃZiNK Q ɝv\4ms)& BpW8d(!X.0WI$]`@F2JmyeVD'/CWIq14J(5[t ,7E:myyvj,Z#%)m]#!pF%&Eՙ@UXh!`N"4=:!P,Y^E4c6F)Uk% aM6YTZN1 늳+ <lQz>i',pۛOn6q?9Ih4IDPQɔQu)N9' 8}t,yrh.gqΖWᡧ QD^QZ c Y\1׻,/)ppeڈ&1bqڨ(PH*+9BOQ\Z9h8P! /\1$dȏk])a4 tjKЭnZ1.YffZ$8e.*)_,KD08G"@fviL|Ф+VqG̵oJ!KF=FղM]O@nmZ0h@Ƌ _*҇ hLQF Jɭ -QezA0񘌔;`j:.%*  LĴyE(C`rT`(~ʢ4SOmh((AJv+5<om -!Lt# V''biV0ۂl2`$ a`1~v?>ndͺ۽q'SaA[wmH4}4 ^{@#Q`=9kƥ4A .p_GעTmn׀R"d$e`y1l rAn! U Qq @pȃ s^ ϠB.ds-Ky=(g@b[-6*8tHȉФ xdt49v_)0ALh4;?'wQp;J<9yCapYTeK'R% f".4C:hˢ9 q,F%k@foj6p*grMKoEX otf&=8]$/6iINQBec0sɚB@`b?X~=NӦ|\Ճnet>+k(3 Zbe -uӭ2o&x љ8 ށfxk]Ь! Zk sѦ@ p*fQUq Z#ӆ D e :rfF8pJf:t H+u!֭FpJ8n{s|d0=ZMf%RrjKچsAgpd~E|?XfP 9$bR "l٥#plR.BFCdw]HA Sh$DiBY~E+Neg,'[:0=+i9ɥq?O mYL.W6wg=? wnII脯/ t_nܰgi{f[s3c #T3'ܢ[?3P{ftcN37"wExb,/]yN{G[>w9𚱸+Ż+{v>a9z`쮞׫㺫'\]= /]'+]ss:vDs'_1 pcqW@k} (KcW]}wU7isZfARjEdog9] jݛ(:]5-N:MïO~j@$? c_k4iZ^wLݛl>f/lOoɌOw~ڼ_q篯:^teM8+t̗:v;WϧF7YHwt.yb1$:MƏrhXGj<5NGڤow(C煪fyyq־qf1j9=,u.s;}LRVNdVn O|V:S9ŧ}=G//VIl\2Y]``5_M N!G+O}+:5RبMCn*ڜR3UrK},3s(Xfb?ux}V(rVMrN])N8iB/~9MIwOgwTw}> s%}3`qD}V:{s.!7A8,/[)g5[qMSΦxwDZђ[ߥ;L m7,FCpU/J[כA%n(Ѫ?t?(39ׄ{ig&^sVU՗7G^M\Wi0ZdK>8(t M4J+&,XUSӿ|V1v=vە繳?PO|sm\~*L+;yƒl;_h@{F)}Zi=6ٌtvr5^Oqמ~+E{lQdeAX}Cwm$W:^ů6`s|A]F͠F;Έ|V>Na ⯹o5ESWlgwHMI܅X&gqoˬL8Let߯f3DG~7?fOq` +G~onrж][X VU'ݹxq?r2|q}y+=v&w"enb®6խl˫7NW? lxއSE7U8reCn}4 IyH麳?^`ep ?9d|]=*3Oz:? Ojʱ-TmEih-CjmR)n{/ rҿ.c90﾿m_G~{ m[|~Dh7^m-Ep*\.DA׎5I:S+OzX6ocGGZ^#)1uC!@ K?k {M^na`6syX]2%ZГL %5J2ϭnMn2rȧUO{OSwЗ(}{esryVܦZyO2'=4^Uv 5ٺhCThkT? *u…9|sU~7x6G (eBJœRgd 37*sv53PpQW|aWYlV^ve;Kz2 mXl;YM`w];{3eE}Ds"7urN:9\'p_"Rl54_y8*Fcq*A6QR$QΕ 1i_Leg7VҤ--:/ k]Tc)-]7Mɞ t9_UʸbI^%xW1+f‰(? ?yx{C^;Կ5R֜ N_5S$s}B,|*φWެj{t:E`T6D*.IP2Ͳ8.'?YG>difyoWKK(G$H^UI<1}*|*R j^ۺɄ=ükbz d45c\V+pSm8K5W$yOhcXƑODQ=rH$Ȃ$)EDIH,Fw`aXE>[>y'g4Y.ɡD"AsJ<9ـ&hd̢8 ;'^~7twIT"9xNgD阋aRsl?aCG>m;ǫ#?#@xaK!!(¿Sd@Fw!x?T{=g\>^-@ۢ >GƁtGlyv\8r6YW bpHSXڃ@/VjӉ*j)O0cV`ƑOyAՅŃ-bp"N2oY cHp8Q'Ȁ'1FI0#2:oȧBn{<\& ~a{vzK~yxlqFI.,`("_j5)6 W.7p /ȪaUBNdHHwu`ҿBO]: rw9gŹJ4ϥjFdX"0Koĺ" -/ؑw131mFKA-QyUA B+-I.Х fmȧ+]C{edݻ-bƄ{IS˪._"M 颐,1uDȉ՞Vwd=EC>u uJWɦaBa`%%b7'\He1*`϶%)CAqR7ۏW#1ۀ8.m;ZC<#eP27n?m=.", _S$r)D\3VQ3r TrP#T܀KvM92`sD뭒57{Kghn,LQ53VެAɮVGo6cpZ0VFګ'XmHɐ纫3yQw!UH8QVĬɴTz N*=ǷKB'|ȧBL3[^v˔"$YrU81l(Zliwqׇ T_o|&pC P]m7tqTJ8s =/cuwq~9O_/(QCg4ӆ tA{X>'T1hP=ztpTP˖kɴl҈]Ixp'`1MeD*˜DdTk_( "x(򉐳X;*f&jK<{"`Fkjy}vEOyk&_\nwo7~SLñ\/͝ɨԕpК?`a3_]Tytެ# lT 6|lxl8@=͖ݲ߸w6IʷP rʛ k;3QDMÈ@`; To_DKi4료N x |*亝0IwZakk5"rd4I\6fD3L΀vj782nCn䌺eg _>f˼'¡/F^@c/C.e-²+ 1b3ӆf'T9W YW(Y֏g]ቦ>>Y=y" tLS:Քoi8KAH0E"9GrTȵm3Oqa ,, k8x-, &NI,z! Jm3r~W~Nt7̑9gE5 /g}9PmAcmc[G؜nۂ5уz!)Mx H*Dem)8I1N΋>!'" i,xqPy2a5ZڅnfGjDG)(z!}XI~4sCxڛ:x^hZ: 4(KD'Zȧ~#3vqjlv&ffذ}T;c e3e ^DX1s7Wyw1?f=tu׷4ˊa thaux:2:H3x4cYIՂ)ӓnk1S;Bs;08 P7$ ލX5cH4fԆL'JρCL ']~.|*RUOw0zʼR Fc"9kīk |ўGF|NY" @=fvaY8"jȥYg]b`'kD;-{Pzr0371 OUbMTRٱ%3[;}>l7:\vv0)߀K^[6~j,㩐k Gg̬=pn/F ŬNd،mrxssK ⎩7/ӈFu@F竛N oT8>cX5[  Doغ(<'#qo2" t9gwemd}od}r;B쵱jPad>kI1d@@B臦PVJ̮GwOI4MۣoNۻpZW*-#PngvtB0FqVM ,&k\ kYQFSmbDF`őOEשt-AjN$@3^ːE,ZDt팺vp+WC:{zl0jaCCPt3߻ScYԠ̛0#(Dy T6޵e-FP-\bn.5ct1\kqp^|-|v4pq:})hFO,7'8r)YКHdr}%-;QF9GrD9gda:Jd!PSA]pO#jz 6@pFΝ(ᣌίCu( .Aᳶ]s s6ě fW5H?ˆY*3Vώ I '".v\]"2G cm`8Cve.U6UΌe-0 _3 >( G=1|lB>I?Cly}V<7\Av Cy毭]qSƖ\2@'rH@]oS▜p6ه!c˵J΁kҝ@s52*g/1P}{|Dĺ~ |"lˬ7fd67)¬')$po6` z{_È5{@Fӱ+ 6Lr4̌.p`9#Y Xa{s`etY3DŜswK?l83LG,`#rGnoVѐsD/3` k g!Fk hY˥/ҕ1FdJb)h\Mժ_y p:*f>m8(:}#RPI2N|dp,~]X𐣚h @h>1ڎn0W9k[tYP4APN匉a>^C+!~ ڼj8J\g!8m>waU/m8}nE>n>&G=~S}]˲#?eRӇ I?^S}/)rwqF}Æ"EIdmnI03MHVˎ$,oQemuږl]4*/_>4 `zz6UV\u4}&n9˼(VpOZX=,i^\E/'y0iP/jd$m\Gk%7ӽ0m.m ;~oL1?2}OJcGZHL]n3GRjc'̝P"SpfRWjAyHA uE~6I)]U/l8?Όr >-3ʉcKȹܩΚ{U:?\8ӧɵg/VE&]e3/UGmKф*껲?]J[S+34_[a qݒ*`YTfn_+ee/.9ϚWeQw}gTQdϝ*.e'u,CGKj;)mr|ϲrlCﲫ;Pq0usYcKarbe]guк}s}F]I5,ixюeqo| eͪjQ^]^R^?%.U ̗շ--" kS )蕝9XAM?;G0]s"T1]B:v<[ᣛ6+ibDc68Es\vF(& 0 |铘q.(;DܾLcnX68DCJ-iob7 멈^ú=9ϧ|,m=>'_!^+C0v589 _ WUՁ<{+<y pcHˆy}(qx!m> 31(dr1>j p'm>'ĸ1Kbp"KZ1!z}09z/x6m}|fž%Fy۱(/ECTٻ$1ZZUe)O=xN-c.T1~)*lН Vh(೻al40 )(r wHr W׻ |Mʧtfi'1+@/~ z!m58 yGO fs` bs'=7VDTAc+:FTҩcxE(kXfnBվCpڢ&`l"nM pJWV<_XF9zd"uե@l"olP97ښ6ycE7Ho|mƉԷlWlB"D# /\%ln#։G89|esd2ʧDZ!TQgq}~x9D~MVߖcXk|0[ ,d6!}@F8k'D55\P)s[j?.M}EżV.evM Gl]r}1>k:`>D;-@̋m0;=4nfnf(Y>թ0s#Ws?`=ؼ#/#nqP&҂v~([Ӊ-XFw~0Ow; ΀: CS9$Oxh p$r RcD$+!{$z$Acٛa]Y²=]wAɼ-,x[X~d;bϙUu[.-J]:bgWQy Wϯ05O{up޼ " !"+ QR"MVZ'h׼<-@2:MYlAp^=/A=/mλϷ9u=wWXY4>Ji,Pfa~TJJnO,V4i`fE48 P_oDTNjt|z$8AY3TnZe@IyKZg S|%;]Q]!9~9bضQ0 ӥ^g!m20 $E3QqxPyG:Mt* kT ePz|AؽxW0d~L6 E膡(Y.$!{=48<o] {izjx`vPViPڽ6e!+ⰃzfwR0*Cy}R>$2$6qp{hnL|C ٲȺn f:UҸD?Hv>Kv颲JE(;Hc0TĈ(݈KKe-#0 gzO /x$9 aN||!U16e!|2d Tzҥ, g3D|F%_B,I`#ϤJ) N4䅰lsP$ h>ےzo+/ =z" ,3dyP&.<h]*+Jt8I^ TAK=Uzץx'GҪ}Pe*mHBYW%V1C(1]I"?,9B#KN}IMpMlBY9H]"PX~Y^Rl*" [n{h|-hV&CwqF ^ʯe+1E[% Z#mM|b*gxL`Wa|})pemm濰Gd% %{ ˵y^ȵQ7// Rn9.t}b E⃩~]쿳TG N+^-ik8XDR[t_;:-lai変S.>礥crkX]Y.op\YFqIZo߼qj֭(;U\fiTj@%9V=fNɁA—` !Kٲ *.GBʎܮ}Ø*@|kCf6`12sFL|0'f1tA?66N+&5Q.+2=~hps-Ƒ"*I$cA-8zh̜T8fvۘ|-9P4'wxҋ,N0V5w(c=Ď"2QD&v݉D*UnDݴG" X Ue(? 1I6ML=qQw*MU- s]9 bd<8QQ ؑDQ,ˆK@Cd\pՀ٨2=4eNpۗ5AP_YW` 1v|D.ESӁ<)% \F0TD<"*5&+N)Ar~0"ӞT-+jKmqԢi8㣛Ge swحUʕ$ppXP?HLH9un'yú«F#pdpHXBE;2 ݍFI᜷hƸq)Jouo1Yw ^8?<`>S6C=!/}3yccyW4e#F2`{}aYV4uZC#Dj,`͑\M5M[*_^*oaj`ܗo_JS[W?Vt:q o5T0-_QMNk*MQ6TvxB,e~/XnX'NBk~}Ϗ*+M`.݂k[ >]_ER+FL4O-i+4Z}"_˖S^Ս52UzWR=Z˶`hsw9 ^XT3H)OJcGZHZ\hE& >ő:ptDZ a*Q>`}Jlte4WU3 C *(I%L|e{qfSi:}*2N6s?3SCW%,",q> L<~[E75yW6s3/UGkKQ껲˼Au mOy~j.}/vV+a3f7y㸍_1P|sd!$/D8[= HɲԶdW.ИeV*Vє߿"-yEJrV`XtOprnL0})n7{)k%j'S 懻/TvpX3y O n NexnVE|W["^0ؓ6ݟqnW&W/_fL٫jaAnƅ-fu}dL? ZE}5޴{ۅkG, ੦G$d> Y[FuwpT~xr߃'柦 h\{lb3Ⱥ{'yGB4}h qc90pG\L>7S$1:ͺP;A3<ŸO}';(㨔aٔ:-|{63tZ2H!exg^`3`9F\#@T$qw^/\pDw<`y q:4^ N3W;hY;bI3by+sz>QW-t7fiKF15q(KnJ|z? א"NG )ʦHt7]̰qG!OQ2Rh]N=]FAFh (ߞ `dQNօT#n[Qm; k]0-I8ʹ@r:A=ssAm-MƳ=q`:A?i,\j&f1I-Y r{|JZʘ{ݍ@G C*7D@Eh*m9zs?mu<`y$C Z~ޱoJ0lH"gH%HF~؞:"wZG x{ LKN\w_cA'H99ONv[I8=u5v(wÓL j] -H=n:fϯ>)4C f(äYIotADLŶ,rV}8 PYdzbv[h$q2RA;,xK+k4 g.11 R*. ,-t`d$A)в`ԝ<.އS S=IM _BfIl:C֝_q:oIFb'>lL˼\O{Xw&QR/Gٓ >qgQGIH!>b!$Sa]CzK4ŜdzNGI6MwI_Cxzt/L4ۦ /s]DGhH"FbEVAM>wwh9{΁8w;,ÓDuG#H4 A|%:VT-Stehzק::<`b<ŲuAZ"r+HvI}pilnTFr[<<q n-r HqMt]wj,H-9X!PQ;vi/ti:m_NPD(i!߾gZ4q2Hg& ˴W6`¨g~:$y[صaq27Xݱr*]t>b/40 H?KNNqPk֟-4zcq2ׇ7G&f22N ^g=]nfzsEn1_Vy| W/NqL.nKƗxb.fTAa4(31=d4c$&CqpehCҌ0!}is6?6CcP i4'Q}~ըES!|t@`;hԛ"8ny{?K=G,^{YAnsҕ83%Yq9PQ'!TFb)OH=ax╘aVV8] o2hd0u Z{psbCsi0@vC>]]r0Ft9 Ui2f/IC>b|Lkh"{m~d!-q!bqJq-цDp*Xf)Vt`#t90lHadD4[T,( eQp$>`ʊ*EO +): _I%oGaB,XcA2w㊷e9&?@IJЭ˨EɃ`^5}IL i_7VΣXNeb#o.rx[n-֐Ryae2AJ,ҙ @ͥuIŮFo.=NFSLi׵~KovuJUm) ޫU’I,r Kv`@М4YҗÖ`cMR:6wT92"[U$8BT.=;l찳e!^!'T).ts*.G\dup.0D./K:p@=Dcuf6ͿRm=9NF Ufy?e=jtɓ&üUEEUӔ:6Ȝ&h$:St ]gAuX4iCS3GĹ>ԡdЙJ, J0/}nuh1慫q =|tFl x?4X.[hdKm$bH4P3ht'# G-f,hvMNǐs4I@Dlq_eWLkt2yȰE>e/%-f`,CBdBwxk@u5|VM뷳@ͼO4 |1c)Tp 7"ɪh_?9y7^Hc8*DU޶><1,q&)~٬6fǷ-Aw^9hxzl*GZr]Slij\.,P9>\m=yh Aړ> I'm+%B-5LefOΌ;n/^=5a\{^L]C5$qaޜB?u&zݔns\̑N댉@,H o{P##sw89&-u5,Y,G0Jɥf0(4 Y3S~?Y(nkW6f"pG_6TApg7u׾FUUX6 my0}oV!sŞV_ͷ~؄_n,{ᛯEmVޕ{z ئLE2}-ˠGϫ_ji}?b |:y~{玪݁qmi@v OmjỘA56،Ӫ^;k?j~T_9;ogyr;\;]mgs`_x٫֗rԋYxgW=1 ;YM8~c.a%b'_[pYn }dQ:߯c +`ؒ")2PMwڜlixrH@HKƎfo*G,zjt ٬V^n~_>1?MS+<ی?xf)+~6(]Y܃vRytM N]xGQQZVގ/|yWHi`Dww8Ь}vyOYRa Jb ަpV1˫Ngiw!d!d)E )˰ /ADt=sh[_V9p]@'H{K69b8e޲Avu *JQ>E,B<}z8WBۊD6gֺ,NO6# Β*RzM4-ݵoH^ݮKW`|G pf#\x{s7MvAW'R"mWW^oJym@=Ve꧘Me meX-|9,8e$X6r#CpK/{MvvL!sv7VF=3A[ܶ$--O2[-X*+:ofmc.@y`Vacv^f׳.x`dwN,㎷`w|)|zԚ_5lz~^n }ut7gbn)uWsiIѮ`(P`aYskͲ䥃5]UD>"3F+Vtuww˼<*`ran5MXDvg`9n׷./O]62t.rkAޯ=6>׳³SffGRlwr\%λXճfa>b}^nka-G|kX\u}tB;5m/׍Xz0F|_r` ɍim}K]׿vts}?NޭV5ot (3wgq?a8?b ofˋټ׮<+v:'V t/<{ryf޴e.3B)ȹQAY+P [t~@\jWU{&1+D3 G.[i޴y&h~4bbδ߬sPwSƂx sg3x3z ,W:q6rTfxw3&`nރ{AӾ3-1I_2h1Q^}*?KS*B ߼BKت,s"7,g9lL)B̍洐Ὴy$dxG(_1#5{ |Z!֏uVɗfBʘ3֕"Lʼnx%9ɉZsU@cF];ʵ=Z7;<]&n{{}=[iec]SlQfW7<Yv߹wަ)u7|FAPQjG e#Gbqe5AFycmU1!Fи v!8HpL"! W)%63hGFT\ksRbhNyi,ruۯUav⠈9o_v3C%#4K+ Y &B%dؘr9T Ce-8aHSM p[A)gGѯ$&u.DT`S=f(\6\* rnܼܿButpToT J0FZph~Yuf<|W+TZ ;0JB栻'ɤEknǒPA#-8BEٜ?~R1|ѴH=(M_Y QtX ~*.$+iQ\f;>lngaA#-8c(An|i8c`d|ke3@W9.JhgVh k)IQ`Y{o3IaPJ l( IJj%_7>vCEE Qz"-E̲ ?+i8?|۪Iܝ;UX륭-7c]W4A}gٲ{hһKA'V;JÉh70#ݔ6`QeA60#W#WwcT'f;HG|Ʋ|uVl*l]L'5nH%g>iA#-8F7Jkq)0I8̻t[Z!%F^Fu S Q.x6h2;*D] 9'xr@_Fy ZD:5L}I448hli?-(\7?Y܄3J0;hGJ;hG! 381bVNxp" >ܴghh eҾ"S(%VQ&%kS[{*/8'i:iqP9T ;RϽĀLw¢_ uT Üd]NyɈ9Q9h:10NJ0E"\)OT[iPfi<ҹrW r_`朔ĦP V\^bJ*8~ZPAr?Ei)/ X2L+M.@gpbAd#_ x\ 1=D $f-4҂2#|#\H >)9h 7ˣA+1^p*\taiq0*QX2 :'Hq} _ f(%*zrcƀ2LNqtH*!4ʹQGߛ.7#ȘY{, v{iqa0%  eɔ*pZ:e$Pɽی mCJJFu;ʑ݇#>8tJGKƀ=tո8H, pxp^l<+8}˱cj2~/nR접itpho?'$i8\] xR)@ ̄)i.+ԸkZ W%F#R '>4 Xn.(tڴWV4҂癃f![#,nֆkO Xq\Q t|l5Gd ! DY֊W eDYNKK  u(Qto/ y|{hvBƆe|%) Ip.tc R15VܾtMdTƠ8B!Eϓ{?I7-$1vlȔbЀ\c[ *q5/A#-8LchAQ#jhC˰:A/2uꍙ_93ֈ`Q&q#:}OJ접!5A G]hrh1Nɐ%Aahtj1FIa1 O;he\ 'ěFZpb\,ڸ~Q~P HFXC+HcpJPyͦEFf3z>E핃 v˓b#A|c;?[k5-0ʜ*>u`tRL֑$ M{K-dpwDG?v x1QrU`L8 *M=~w`M2^v~Zer&<6Lmٺ͝mNS WzmmgMB >%kXrB>Ns K׳vy~\ۯ*l7D}-;u=k.d{}^uDM{^̝vZf٧u] ao{O.h?2xp1/΀Yy @4o~d:x ^}u2_IբUg s^J qV`Yua*Og[F4AS|daC!|:FGs=Ս ?ByB UֶJ-ȕsuJAV`i[麾0JW k,{,{3jKߜ>g8,g:gtafUڭ(YjG#,y ٽT:yc'n3JmG3k/]Iy0y\(fu&љ|w{_{;N|:%3)tǛ¢؛Y兑=8ۜesnYN* ,'p h]KMY1h(m.ej&rbp%ufv(6C-$WAofRRrlfRѯٟ$ Wý+/]\tpq{+KW0_՟G\iz,}Y>⺝ק?ͮ8jՙG17dA<K9rea8WV?{FL:xn&u ^Λva]n& NηIp׵=~lJnc:WofJ/\g '[wٿuĽ0!~F$:K4{=&\MZ8~AN}P%5X-Oo]?{؍dlh0v3F|%G }W_ے.mmMcHWu)>On/a9'B[@}uNF׬58o'KV[l7l!MviEײn)n-W- 磗?7r=8q!@w[PϞ/GghFq Bg66荾Cm .cwiRZX" q rsTԨ/x姖;׵.d WJymW}/'>YOF?2/tps녀5rcPɴ `E+-lĻ{MFXܫu+tP,1 QB9 Qh9 u^P>ߑ]ġ]߲wYf=-;U͐rul S. &hNoSv5ϖYϳwڰ7rE 48"FTۘp}FI~W{N7ŒΝC8}oae-{F-p>y KJ^m[+=C<)cqz!WQ.rimKk~i풆ܷ^ i;7d x%O?E刺n4iLH8݊!ǨE@Iq̩,%ZqF NK$ s ep1AqpQd5sM`ApJ)gI9u˕22=BHl7TfчߤQB w1/dX;)ФeZ$̒w!F6Xf&ό{1iGΎLsj%aw]#"y\>XwrNyqU=o^ls%t\?%Z~ϯ߅+ISQkxLc/|5 {=t[n1gǰN/ް1G=t7Mb6QjOeܬ&lS[s=Aowghrr|/]yZoSNz5,OzϏ0R!OyY0 `pĒ\T¿3&{̈I`ou]W}e>,F|VCZt9¬!:+~v?!x`wυ֋.nĺBt-Ft v.].5߂/ϷgLNl>y7/3ΎT9 nQw9norJ[W=vU7kOÇwN:O#`m&Gx\xl+]1/{{~>Dw:˔OF2sZ+˾u+t:_u0GN,NW^-\ο!xAӳi;,y'FY fݲ˧8vHuK3|wM>[F&ꄒNy%0WoӦ6;{IRQɮF Ib+w:Ϡ+Wbtt9殥9~^Rl2b`Xfy ]q?6^aϕvEd30x06fk2?>h\`tȝpŤvvo~:|Gn 7l'7Bv^=|~0bp;$iQ.I u>:jKktjeTB4 >|cAC m3O}'uh;_N8'#{?,*Bgn5 B:Fɕ0:z&|BI <[C"s&aLDO,hn0K{[-w{གྷ<ߏ]2-آf]f3|L'2X^; LپZuσ4ås5cBʘNBk'2ӝ\x=q$:TIYnځpvF3?v|Hdg>;Rဇ=G:yoD⫟T8q>C}NreF. !!)xʹj|Y]7~1"vУq-f}N遝Oi=b=˽}V]FZ0!ӫHP>bqMM֫ FԦ/ɵ=rb}[\ۗP nI>573"e|SX &J Ѧo?s/诿Z$6^!K$̓ *>Z5F-[H WVSqu3ɩ`z6Az^#"#4FisMPok j:qq3Ey,@O8G^ 渵FR&e{T^K+|?ڲpΑBT"5qNNYCR0m:oaCH2$!V͉A9u,HxJ<&F &aDrf耥ǹ6sbTSg8'&db˃ͤz7iw@{F~ ʝ@E֨k m]tI*oAW ZՂt]-j]-jAW &R#H?#H?Ҙ0-M,LPPW`r(BT>dYK[/0Y(ebVKڷc up5ڭv } [enuvū-PW};{maONS}w֒rWYaI]eu;w5䩻+RvKtWB0㮲B*UecW_\-]?{Pj/?yn2F@#XDwqrݚJ\89Z ,޵57nc˦wB!ݻUd*KMҲݝ LISؒ8=J7 Do~M| dy,F'b^fHh/,b̘eۜHJ(#yP(% Rm=7 g#=]kWZ^79 oM {`iƌ^817jiwL.U{([kk]>Man0QSKւt&DFtJnP<[;nec =P sZF^6ղ}4:`ӧ^OG;צ 5{-4o։&Hhap.)I0CwFPC 4!ѭ!̀IYfr=m:ZjNɬgsJF,i[mt'"! @^F&ISy#:> -^"|:MRO7 ˝B%k5<=㱞ELhh++DGbl1-17\&\P`.<)\d;NB?;zٖ2(bS|ʼngQ/6bPHg'rC}sMͳ;:|~Z^eW{`GՒf:K1;%޵X*`Z2S9r֕vyO LJ&[dO)2_n ii T>wߏ|Nx;R!Y0ykahM3vGN,'2@u"ى9Z_9RoGpP0<$$t p9^<ĞD`L!ʞcR"a D*1So"[+ao U9-A^X?x~{/'"B$8DŽ  S8A4|@}_xcyc0f$Pb 8ii,)]xh+ aVjkmgUrG^F*ģ/_`1ח"2}UoIpwW@>ߎvx)^$R9|NOp[5M󿉗i6S˰eʁ`[c2T9/^ &DD( מFi߃R kk ",M` 0!y4\l+ӕFٷ]uBWEt {J%BW-}+9'Z])WkmOWe\ڴ_ lOtRkrW-F)@WHWLՍ!2{<$9p@wCp"|c$do${郛"Qſ;[=.ƌ ߋ\ͩ+2~tqFu!n}/QqgF -4)v9O2wPa\,=Wx6(y0Ґ1hƓ7!ʔ.ξ7Lo2qsA~d9&͖jl֦l24dʜAd)3U8j1*{TGIMK{8لDe*aiIT6naDCn! Yc YKn-Nh[KL<Ĥ\]^-&2/B4'o vPRO2FxЏ9~o)C,x[^x@,;U~ .tcAa4r:+eWePXk\ P` N+p6D䶕]ihQYiEP؇Ky3.xKВ3]o g\pTC{Jeh;]) Jfp4ZINW %|++j1Ka ])B[JŰtQ<a + X])[JŨtQR4eT@j]ig0\ Ik4ʁڴd4\l+V~cF0`]i[CW.&ծާ5J1䮮qJf [J=W(@WWHW)ӕ =;.ЕF)tQfpNُVLy*.+ey"JޯT;twSTd&Ҁ4\l+xtQB15ҕ`J>Ap;}])ѕBI]]#]ٔ@r #%51NT͡}P: 4 ߠYk*TS~jP} 4} "l{iћ)i.Ɉ߷:ZFySbAm6W(Oh4L$JJ󧩷z'?ͻ?Ng OџOQ:E|SMK}|4}v2S()OS={MBvi.liD%Jk׈yxa񦮊rq TjZNe|p$5͓WXn󯄾igmW7&@ B[rQ0hΌ'E?Y9) ,-%W@vf6GBjW-E:_ QO Z -qSoC ,<{L3)oq+6yE>XKl$8WBڵ& Z׍Mi@ ō-ԏuK8 eq|5sHt Y/r3I_F!D;P@2G+U0 HDG[pЭWR[D͞/bN h1BZO렙W>Gh$tru$XL;5ps)8nh O!Wo[?PQy$Hb#Όtj 1Jz }`r绘w/i@@ǔ.5\].R1qg\bzHNh-&j"?.MXCGi50{N^&ipy8Iݹ>2?THzF.aj2(tcy;& k 4] y ,[vL9' n%!îc *ny+  1iT7[ մAiwH* ݧ$Mޡl< cy}Sȣ0G%sSqؘS $p#t J3u UӞܥ0ۺq[8m3mYL^Be'Pޅ5ڀLX.Dƺ k8|n1RǞ>p'Y.qcXgs0R˞>p.y B{JS A!@I%WrzF5{?WPIOdV$^5Ip,cU/L@ndnXۍ[)s7"wk1X}Z+w-e*all>N+iosûN-' {mjc%擘Fm0Qy4$2Knti=b/Tc d%mL.ˆmY0=uhO3JYVkuvY;$;f,4zYM!@#@M#Yg"Y0v'q nlъk@lZB/@05' Zynvy^U1M#V]m$vL% t:u:1֭'t^?KE^q+l?-W^QO\mOsR g+ߗ5h4q!VpvC$z^|vfnlV`6+۴}*J_4_ZpUljmsl¸Qz|g\9;w%Ԉ`ovpuu@WtdJެ-vĻ]\dX"$ηyQxOC,|UJ6FTfP!e,6"#ŗB]YH\☹қņmi*zeC'޽vfhd&a4ۚUHȲăzK*z˘;RF͟2ON\k\ V_U'~Db_\XjHxIbl3{H<5 ɥ}ǧS'5rFYN''0~.SxjH$Ë IVHf p kL8;SuẼt];}efqN٧\I7{m_L Y_Jrt_)#"JW#>cڹ1#G+(*qe~Di3W$$9f3sK~)}<4z}VqEcX&uL^h]Ɠ aD P1.ƥƣ"/ D)3@ifGNLvz&>G)90_q:qK6x6II$"҃KD `/s朗KG<0EsWB"TÇ~N&QeT'rWwD<Ǎ^T"-ΐGG' J0B_(*hZu2K?IN% *t5{c[c(qf))WP:/ WtB$B<!Fh\ߜBbn&~1fRT{XL>yaTL'>Ho}V)DB=1bmec .aԖRsa\aEYS#Y^}P{~6ޫ.eɖI13]tDsV ~n򬙕gkd(XvfjTؙpZ _ͥs-1iNN:"t:ch31ch3͌(g2I\8w@443K] V͐UDOn@6wr+vO2?ywoIo7k/~[0M[9h#B 6Ý ^%6܆VR'b\B @܇aa ZQ_n975WwJ'4UiۡGKjL$E/Vi|r\0`/lhwm&sպc:2Xj{Ʌ5kudFF}"nE4:sQL(E21o\DDhy@)QLOoHP;Jo?_ 7K/ܭʼ_ù}˽MqN`E^4$L {<#De/\{OC^WEfqGB`>#uL>Ƙo !ljgNIg3U`?8eld^peRʇfZ#:j;[VpNx旞8^%F={)7.gK6riʰ1).,%(U8Ho=#K-'jwI'sÂͲ)38:8 nU}u,2hwakand&G>aCiRr.lfY]C *k4b|P?UHmS[q/ 4XV4:t,ۂe4wV'%ȕZrVZ^fx;"k9/76ei%R tP 6ei, Lcy$:,$e9@vn_)=&dRÐK9qTJ<9S逰k,e hC;PV#Gfzx&1M4KjJ`|I5vS E5A-f颴y߁GG(B�mq܂¹޿pϫ3뙿6hv:0Kizo5Sue񏮸a d8¸ ~ >k0cvR⎂{p H#h;C/v_0uh"שq$Y4Nq8%/wFTsz|RHPD{B)I NU>C_xPR%!*i}+<\QP8S] r|PJKW>PD]?B䄫} ;gt/}pYAGm]e@qmߥR!k n (Q^;=jI VrǤiuV%GtV@}yjd}Lqila4KW +elU6d ?H`-bJ# RuNC6me= DU`-2# Q=>oPQ:I=sef.m!mK !ş/;msVXƻYayڵMXSq2Ƃ+#W `5  .Xp$JZ*˄ v*pG\¦r$UX<=UVі*2am{y=ЦtN[R9n*a0ZI)H$!q-|ER2PEJVɗ mۺeѤIn+Vƶbel+VƶbeؖI=v[#A3>; RU`rX qWi {8T$vLnLJv.]v(,H Ɓsokp Jخ[pVT",1`.zUE?.v͵Xx+͕ T=hoS` {kR0 Xp`-E6&N0l9hdN56  T[`MRvوBf9y_U9w1FJ ˜1 SZS(DO?o$i\=.JHm0 T{,gUrh{A\y` lgfI{s=/:((}?TXջ:zXGï\?k&7hz;Ay=ți|~=؊K8m:/VTx<\]+67P?L&||w[Q?rd<t"b3?qwHE(T> gWR1rX,ڥ3X8\kO`{XGAy hrc1rcO+T=F(%8̾qϐ12w7cFp RtW )k,!I(z-KX5% -%I,޴TQ#XLT&BY%RI={nɾcb.E!ļ!cHUJ4W;\ ahQ!qb7(G""QIi)t\:;U.F5sA/5$bC.W=^LSoމd1'W=Ƕ,|Hj[67Ě8oHSMD)lQR1 wކ/ $ކ({ s27;"cn&7WN&"e^Q  %Vz~eHƪ03uC{aXאBlvwBơқS{OBj܆$2RRY!erʽSb&:]2l֔H~,au9@byA*$gvLQ|AkmN<iUNvg4Nh鏪9ځe?.G(>%^|PJQ,g̓7Q>AVxS8O !xтɀ>G,;L,VDj 'rMv,Az#r]VD^ʥ$lLHa!ùBz!a=@', yb%7]T,8KH?V<X]њNOU|a~Ӵ _cn(yb!lmI~ܽDE;Lͯ+jZ_B"mYRb=}@Ŭ3B>n+>l/"}D$EyG1h%va ⼠DIo8ƼAB΄.8FR J'70C# 9#J:>|%e^Jy! KWmWL2R]xN@ƯgdN璻&N zK瘔BJZ1j /q_abg*j5!o)05,%ԢbfPa:{{P ̓*{t,gL}y3|s3v5^xlPVvJiF!{V:wSBqg^{1Ch4݇GOAoCZ;n jYh'H%Va'SUD]cD⁈s(8cefl T4f42RߍdU b~`BTG"muI0*C#r݌DyMF JNmI]q:VBd:ρ)n];_fTetm(|0egސpo7erRLC`bxEu2[Av柹߲c$(_PKL*64 BK4[8ͺ=I7 RYD0}!Q=!!H/zrCu_9_sxc|b^I 'G@)KTjϴYWb3S982Tx.G:{*{;btd'`$gSs#05WgX ^ ΦfYLv2B'~ID )VǢ_g!F@jQy;p[ms $АEطEFB?y=[[g.M((mK`"yK_M~W s>^j??=0lHΥ|;ٺe^=;;@ p63"GrAx=^r#_=%!s`8 d}DB$K9yg1z|IQaZe&y{ a94Y "-?ܒnGOtK}k(imQtqdoi-`x lٻ(mY&~M11˹msp, H>[r=7@DDtZc{jH:YNEOؽLsH5Y8op4VQ&"qBI/qfLZ- L Z(ThX[s/4 DY&LɅi"|Z 0( -{md0@uy[t>J1)\eȫmGk]ƝJWé Pd`ʫi;A;Bkl-c@\$& 6逢@0k*w^A!pIRA\ ?b`H)1fNA͊zPb$E>\䐗+7s8=Uo;`.Pyd/ww p!ur|s0Q0 "W Eoq-kH QLWr=|"t Sbau=@B>xwή)oJNnFSi 6,TߔB|ϳv? 7 HA>bACȣ%O(9cl@4Sݣti%Y.Lֹ0U=oxűI}=)R} q%KUFaU^gÀJo^*Bv' [mOkT^/)$C`be;'IPR2Rޣ=Tk}\QeK:RJr=IX3w#a#^>*zJڶW_w3KΆc+¼VX(Kب #1j{] 'ib횂Hcnshe-'qb2*+*^IAW(C|D͒FQ ջH&"5~#\+WLRҳjG)P%I؃7\"9˴ H{!*G̿&AD,|^Aޘ$߰M8cbeVenf Ad.rM녑~LkPy6[m2|@]@ wlK-zO|ݕ{mR '*oǓ PΆ:,:g:)_@ sAȺ7nX!ٸԟj$*.Լ1 Ltd/ ld/9fF.>]'qsI ¢iFL \4/)t",KR~P4zOfRRou`H$.P@*{ʘ ЋzqМ 8**9K :{"C\J_J$=U3A &JD\L[8(t}*EA~3MԤDz.pƺ%e< Tyz7?L 4WVrv~ɵ;DI)2 @{p(8W'y7on -SH /"0_siATߍs8xQُEU`J~98=-` j)4@B6'/0I))vR+ujybReQV{QgG7^ZYt3u$F4sW;~} ;1'K)جĖ/fR_fayp@9D=$4ACHo~TTlŬ51L>?m :iaE1k\EJ}4-AQnۛ#~z66fʣb Hj8/èСuY̟yxkX?H҄Q@DJr ж;۲r#ΣUZxC欲? h900 P8TT2ÐhSaNކGP!X|w/>!-8G 5P AE3i-k1qt 16n)a&hk{x6b~R&WQ^Q1P "( 7` q}m 1k46Il@\H}Y1өҲB `êAA+(D wE Ha-fF 8#JA~TX)8U|K]SoB{r E"'GF{LkEbn96G#{E~-k\( 4`aC(P\,SvF  TvYbj#aiG:J_?ϐB݂|um=ʤ Deaj3t#SÚL3{At wTZܡ<^PhE“ƨbfnJctrƭhR5Ƈj~{SS9HЮDy/)W$g(1Qgi;Vv|/I#z$P7$)m\T8c9˝YpB%CC)kmF8·x-p(2S vàX8eGR~);۲)eY؉W/-:vhg3*vV~)|v>kvΊ잇-\!ZtWeXs['8k⚝YEOٜ0(#'7ݞ_E:EcӮaax˰&Qܢl[K-mXc ʩgӱYY>@a k\:WP_腆}DUxX!WGe =5Ue`G0I>TJU:쯝b,$%e˨βsޅj]'P>Xpn#a.}S4/VQ_C]T`XT vTrRô%agzvLIHdk4=Me2+ *YR/oBJsh!S3NÎ`arNiv)-4Çt02o-<:xq o[fwumuTr봜 k\.c*zZ0מH-O䠜4-edcB^jW[Q%l:w~5;_]fvG#{(IQm>enmJ{]+'dwk=ֈtv +èKvsfL+ﰡui.so7TXWI c$pa<_z-=}&)$S iy~CM.)eh>IԒ K)'Yk1t|VaG9)_՟qF#oYb9_Mc?*~~߻.P!$6.)Υ;>Mh&ɻfbKzKRgYyC4y̫vfBڭ:ՠ0ޮO50(:@64#.ڊYqhsNGʈfIB~,jP_Ϣ`amC?1O~Z2LI?OY<ݞ,#ډ%VQ5hpɂ@@$ervZ _ahB}vUTH ܫE@b| ȓf7g糸T=oj]w e< nw qOZWTsf/GjZ=x#8Ox%0[fDΙS#dBsB&I iݭl:! 8$\5D=ShB SRJr0{z͌9+8,ZW3P>V:M5TG }k@w!ImvtA&ekqC( emb#Lyvv{~8f8 زz^w,S??)\A4|##VREHGLL hQ_4CӀاFq) X J*C!URQ3NgofHnBX`B +G`2!^m[Sŏ47B 3.Ť2̗A4ڪ<5[l5ˈ dZZtNȹnsL(ĢK+aa lFH&&;SOы֩=~!ɀ"VocC,Z0y!װɄKsn`G1jW~DpTo$ =;/7_~sFKv]x'(YF,n$eS-V7A")ZVlJ9u-X' rϾ-tG&rq%F/  GNm3$H":!eV)v`v/)гx^ߓls"=6ٳ$fUOGT繘/4xgzW? qβ1qSP!l$a$ XG =*ضez1h!̦tҰLT)@2ٳ0;AMaLHܢbI“ٴ&Z+,:T2aیP/.!F<y4gF!bV9kR)C(1kTGyJv ;9qiBlr8[ !3R3awfc wcND G3tC̃w82Ȭ HLX TLD{6fl8=$BDǒ`AMgDNF1K 0A#Ev$ sZΪ{i ZXme/RgΡqƗ1878Ntw1$4y O~D@(r&H^ԠEV]OgmM&j\Rn1&΃86*ݎ,1U7p(+;}߽mXc q8at>8Ѯ8;Ϻ(*Ay.n }.ϼHD)ꠋ@~J/88 @iB3K rg(4]ߢB?DK}@ziO6Z8 ,cP-6 ZgnAU YPu\N=;|"q^y\WoqrR<*)Gc]3βa P/۰ƅ;f)=2t>%z]8qWʼnڇ0R\}kYo {64{XiCJ}u̱0&X5tVkN 6iecW].5Ln(;T YMyȾ^#˵d]y.-{-ָNǮݷ(da8A.Qs8iݱ\M&xCW#0k SΜ{GU{}aB'PWweJfd k *8봱^觶א)3NZ>wj?i`46 7Q^Sd]OųdjXc>'){tϴ4z @hQ_/[-$.cSݷ!doAwTC WHPfW[VWDP'h{F]pHZSé,&Z=M$@n |Kr0*Q,a ?ar4`g=EZgh-滛 TZ^ܪ;P3z=V-K kgY2v.AYTٵS kӞH%M(nqAPhXc وB6+ˇu%:aͯ<$a= hq:%糿gWhdKǎc5up+Q􌣣.77RڟD/ .;OH?Emح&RlK՟qob}f?6)yΊǿxD N(ᔉ0))Rh=]\K"܌lq+\AdqTBwij Ҷn0&Ŝf̞40?gU5 QHV2bX||S 3KP7۶NFh\sՏ1ujlE=we^*$=1i&u+ur"k=ɾ!vɉtKq,{oK6"aY9Ҫrq:{3e9JϨԌjO*%t1_,iϰnWeō?LK,ų0sۨZn 8@;S?ׇ?8:lz|!.sv5aW4u},rml)(u1Z.i8mՒLAXzT- |zWTSLڳƦm^ J/_oˠXf4*ATg7E\3JN|Σ%̚tLn,)@{"[{JO9 < P_nv +a5=Ko^cf?&8TƑf=爴0 JMÜvS[!$ijyK\Ǐ#POA@d칪gq3O-'A%c|[<|T+M\Q{VNuGN[FKQVHUcnpҵq}Ղz;[pj_Cs{}"1Mo\wa TzN]?'%jStS>!E&IOA InmE80<fFDH:b-˿xVE8nS׎YZ3NܔiTT0 KirZ2ʖ!g2E ÇA$>?HZ%^!zqN=SZm .s}!\׸W|u; Fץ?{0QD@i׉1PW]` GA@G;jrJDM!I(DgHq/njޗ]aUT9%;lE$|mh|3ĺjDXTɇü47aTp'n{aw)?}FW2=l5f;풧ϴG#bb(WKu72~=oy O~R݊ʋ(؋t1m,` BrY<:kkڻ2iVw#aZZT νs8Bii&=˯"zDcT7 mJ^($aO{^1b.12aX2wJ!c5܅yNƆCnai0HCtZ%+eaLP=.&L6}KW 3&]Uz()5FX)vWv>+A~XtH clp  tмM+H50 C[cvDp\s =vi c[9BYhў IVZ|)BT[G[a f4֧ۤYQi..0{WD]k1hlZP:sl/N y p"8$n,{wQ)"m#ќbh7ŎH G uwA'jM1nn^󡯍Pܞ7 LHb>d3¤';odcՔs[wC!w?Y\.S?l *9P;2J:x%8cԑ]R.Bş?cQ?OWH@9}r򍼢Yqk fk{,J L8.> cVEDV_ mCD&n7_ 41 AzǸX $9_A! YLxR.&37i~e*寓lQNtO׉y_?i\?T4+3 [R/nMbPqIڸJq_P^=?__|Z%CXghYgJg5'ZcxhXKW`,pҲ {9?Z:M5L7NY?pH[VYaX0j-o3d=I㱜+bR(,^VT!&A~|̢_cEƫ?"w Y\>/mP\6,n>>S-۸/ġ15b֊\]>(WnE-i IBB%fֈ0T3uPʈ$ /SRxb34k&j\Lf&W11قu 4OR as*{L])$8;a^`1a2=Ez&LU$-,ےڲC\$KUR`C#ύUf}?qVPukƓhES3V\ӽ/ fkٔ *o ,am0qHځMI~}lMr)p;ō&y٩1w)ṷlF҇j%ЌZXw33cZzܚD`ӰI䓫$\U1ia< " 6o2&~ь*VQ4&hI {6+\> NJ'R L'޸ r4208T(B +8ƘSTAcI${n E*3\\>‘[E*]EϢFZ>'sZhnBI79Jdahx'9*|sV1[ʣegW9zuz}soq_΄9NÉ sȲw{: ztA%lUg?#f1u⬵sTAf)( =a *~XABNp;X|6&4{ D X$|qa ^ f~'ƒˡZ&?N(6 WA3]Ϣ. 1I`)ءIdʬ^(aes(g-zN0N'|~4u<Ö Gn87)O/dm2 ~ Φ95`2`cn쭮Ly"۔1O@d4aH`ؠLm*h.۴JZe"+"{[z]E0Q, Ӈ @S%iz޺뾥5 vKVԈ̪Ŵl[_4>j8 3u8k߳׎H1,Y`h#ԁ< No&IЬ}_,{-}b'C J֘/`LQERM`3XxҌ9EJN\=¸,3Fs"& %4|M1;Ϧ`ʮ;$Q #T$'2|߫(pPYR4;aSaNʺCGiAcpUIfks*QΨLY#_7cļ/xwxu;8)r^ڔkıp`9}L+Ͱ@팾Blś/ rzw_ATJ p\#"'p$D#P *9< ScYcuw?dl?M%ʸs!P pbpo#ٌgh6n&2u&bh2~oA1ԤP!#MM(Ȅ&2TxM:|⾌Px"^ ^CT7 ޝm'o4#Pa)\YG6ʴ/gf`Jd+@RaԞ#qqLb9,2$BEm4GJzzd-ȍ ]e=Ho1~Mw@†O)Ͷ@S䄏?)Okx!4 bD΀e Zf-i`sQ~ u3C(Di-ޘj*(+}b)HOe)0y(Xj_ޛdI\8)J1ݫ7xj3 ?kyS..7e4pQqc>D`'"ЪҶ9t?m5m9{!#X涭vaF3Ct1sָFVqퟃE8f%˞$'Ogk:1'4zv"0'YQ ѭ ߀VN1 xDr72qTe;є=p n#O*hSdX,{Dp狂|]l{>؏u4z| ctը;ݪc7chd!!8%dl>opLsmcؾLP}osQl} gn#TfBzc2/Mi?cm1I-q?MIr;܇:za(wi8Fa>J*X N &rP~ՏQ Uuw ÿ wqQӠXgg *̿fc%`O9$+;.ۦЅf쓩e9~ ?O~˜ ~Ah38yךyɏp4: bu~ ~A'fRB#;J{X˩!á)dER,χ*λ`@AF0z-z'R`Lp$”/pI#Ih 2`d)+F:;Y-1 ±|W)!$0fOƟgLrr@Yӂrs >(5&9UtޫWE6 sLQxȺ04()IvG4ޫd&%R[!(.uWxc5x>̼}DZqF:Ds7}\1}3Cb8e~:%XD8ߧ|K˦3Keg?"1vl3l̥[QS>ODu+h2 W;TD!@s?ĵsƌAQVV&B*vPWfF΢䜨h9] xPpPSeD lJ%YjoUP3a(/L7hFb*ԛ '7HKF%6^hiBB#H*'pR%zwI$(zIci>YgFϊϰ1B ^ [ib͞BXK\*~m ',>^lWz:O\L"5hs^&BJn% / "NC:)&Đ׵L38{0JT߉`9E°4Ӄ ۗ߁A _jy.\hHX{2Iuz7cIcZCF%3Ÿ<kz,5z[gրq+D oBذAK=[Q#-VհV_.|[3wj>x] ~8ћj1oc:BԪyߨɳ| u"!@$ 8\d"d ‘4+QN.j1^^jp]6<Vg~wrYMZdQ붲!V4u.JT DvSsQ(,$*B:5V8bZXK(Ob3C(DiB }^j*͐9X><:E"Up$'Li d&*h9[B\Zsܑ#v+Xo(;M’RskN0+%Yh,@mlճ\wFM|ntg:u{%YkixfYKYJe?Qcq?} LǺ'8>E.~J(&85inFsIw7̋ʸ94z3cH|u6cӲ#9):J,M"[Cfe͡{p#gSA"BٖK D0rFթU_JvN ~DiCm@Km;fAaLCc_⹩Q9;UN?.*EZ-G@L2/oz.eIM =}dYMaWAd__bv /IJ`QevIE#m(h-PҖ2S*N><5i.ȢRY/WeG;z?:f*h"j`]g5"X_ i*P!*pa'̗3SKT\c` sQ3mji\qgZIIM#͇j~8FU?6fEc YJEFnl2"2#t5*&~s_97(-G %q0R0EA+}Wږ, h6V:)^FNxV;wCHvz"PCsQTEG .ˎ!54Q#6%ƮX*/K5 w<nVc M;UgΓCé\'Ib{гv- m3Ll+t ߚLU`?ذe"y{`y>d4s3j"ŌQ2_*J9>YlRr̆ң!90C6Q%-¡w`wJ b o(уK \vm!Cd>G/L]ziKr!CnW|ԳeHY8]%djh)f;tz^{sp(;%'3CL W@c0n^M4Lug5M/FF#zǮRu>ьA]~o0&2G~ O+c7!-[>M6fq1ZuBMOĕӻ{kbrR_O[I[RޝQWWstJ׶ahgЉkIvZ<%A,A]iؔQk%,dA\ټQ:0xh*Z@+I G~7TwZB]`Z 6]М.?v7!^Z8zIQ`F$L0q"Mz_=~P)U 3F .Ĵyw RȋFF/8RLׂ DPHQV8 jhld4-)#J(DUQC)^KSi!T%.0=h W[ Ppl3+̤U߂( 3j+Ie쀅n$`}鋢r42A±B +8ؗۏ+ߎ5! |+|'0nZB`5JGbov;M)WCdE=g+),dRm`a@<">FŬT)lugV1nb~ hR+q TJR[;vK_pX8- pĞzVE8Ap%,(ӄZgհsvq˥$ 6Ku(k(na.^9<)ӗ? '!-8xAl^W[ UC oQJz 5g[%" ޙʇ`|aU:m˂yZV()JNAH[B+-(aes\(g`&$¥ noцRסlh?F3dC:$m_D'A;!$ \Tn02xլ,XuKAO qP)>ʔ~$b q$A%1"MH$g;o} Lf`җk gNL<Ք?Gчb`,W)Pܻ"0v =D+Ju$2hy2F GX_5upunkUn}iMY:Z ^6ɹ%cϝ F%d54>}D@{!+_7hVDmvRJFSKl1;kW#c@MzEɴGdĔF!bIFMWrRJqF檪-S^yvHY _$S{og%h`VJ-ԀKETU O^c֒㊸R(o<*v;!estYz/', >1v^fɜ%. '`dh~swד~VZ醑P]c##lkA/;  NJg]LVL8"l$\Al-HI Rc#RrIm^,IZڐ|UdtCZDƹ"D󋳇(o&nԠ9 |_ɚ"?Z#QL6VY.M$oƧp^0$ѝ$ٟ7xh/o,R)Lj&2>4o.VQbXׂg EfǁǡeTҮUx j3+h6RT19`z9NeA)a V$(Ȃ2TCR?*}u3F"z;^mX:"֦_O0^ %XĻ=}4=HJ-4=kjIZi6[|ҤK})e a0y(XKTC%^)8f%ՋgN~zJ7r/%1:!3C|``lEJe[TI.)/sSbjh|25OiPT+ĕӻ[ѓ?oPvἬ.ȓ"v936 0ߤ{/BؼX*D_%7“e'?ۏ+>Yg|ta柨˖#_RDŽ?X'z;" ѥ93R1uP lEkgPog2H?RT?pp"jٶ~:uQ,n'q7܉_Qo͕~"X\Q㭘.O0%^7TRC( 6,~9oЭi8N9Rp+fH)uZ`Z͹5¯_VO W2YzG<=R<%BJDU1y:sM kW. 7Nh9scFÃoSA伕'gԵ5Hwѓ./*UW.Mu0:*,̛^OcβcwuґƧL Vwlq]/쎽_CT0&楙և&*vB&{X{'P:t=2 /݂Xn7ԐY,xjdb˝:65L!꣔ܑ{EeHJz OPAo$\JNkWtҷbAV߶\g31kKSCr Tį%twfj?69ӧl* iCns$Ø }M Yj zq7y%ύYsJau XkP2O7E6VּuHF* u$ٿ"ikgU?%$`^fAƱl+"{2)ɢ(ʹԥɺ+$ž]]]uNWWjz6DUĦ tdmf팄lreGJVk*٬ :Ф<]%{6Gc{sU$1^dQ}'ILd*2_>xoZvy qhs:ygg3OZQsgooN5|m6-c<7GG1Sn"&^>sO32-c|}Pӏ}}_En|j-.W7-c9h-|Euyk'{=1.̉[0 w[ghO~wᅻ1IxYydļfE_ MǴ{ s{̈gs`S^,)2f U; F Oy)闵Xؚi%\R)'NpE̦)ߤ6WK=\d!ҍwUbQJ%e"yҖ\Q\F_:ФSG >Egy\]^@t .6 n*z0i_jAYz=WoRʕ֪ec,Bq\ 75Zѧ4Iѧ~gMF;. /zE/7×ZSN&cI-[oH 3} Ii](Ŧ7R6Y@SSN/"^qы8.zfM gcGUM,ڐdcԹkAg^ D~vLhvLN2|($,o\cIkHE٦!<,K94ǒv71sk~g-: 1nڼ`|7jz9̭Ϸ̜@*}޴r\u<#0op(ǻ_^\h\\[Qx|{bTC<'7o|WLeww)ڠ*=\Pí&d[(m&`KZ|,XݠU=F.҂ΩC"Υfs[8Z1EmlU[,ku> 6|q%dLQ9(-&h5AgM|\z>yv+R2o.7[EgV: {t&R:V-1>I\y RꅠRe_:a{$r{M߯FbjK5-\Ӫ0JTB(BŗR6f|C=Nr٧˯9'Gsh[$# 4\vH>ġ7}|4FƠ6gX%. z1I ݢaڬ z(Az]أTod ȕj{<BUD9,薣Y'ˤ- &A(LX;~eU' Jك e0wt;5`Iizg*2Wo̔&yKUM e6ndJn&1֑lX!_[)VHjRغwnoSTIp8yb䖇ѵyxMid ۺͫf!A>\IWo_-˷/,AbY%3NCbI+j~`pJ N+ViH9\ 9EP2ѩ!S;Wo@U^)"wͺ<;5OІf^HDwS}E9I1^;TBv{nv[)CzY.[XirpdhvQl?|i_N~{{M9:ps(xXf|Zȗ~:Ex\!r\(68Nb3arzkw+wXaRߗ`Y._xi έåEkK?ZN/?L=9OV?GU<9ZӣϬ1ۚW~g_v>޵!+z9Iʎ\X^/aW?^x|;hV۾<9XOsgz:`m\V2ej)SCOz԰25*b61hq}ɭʮv ۠Ml=JY;8CRС2 ql:#sXJ@u֥$u[ lH6m?H\D|M|I->Bo g7 KFו1մ 1Am[ }Q#&43O@41`=MiA50yCd; ۏ1kѭzPI2&] ~9@gU 0Qj`xؒ_vTP tI*bx"^E)gV%᎒W]@6XXN$ق/D !'8^M-_J- A{\VU衙(B~s*qR%銥``Z]stA!H/G3-(uѓy=8ٓy x,Ds"H؁~Q!7DaOad3w7U!P?]U}X2faimZ>Ԯiھ_K_7>pw~&͍l,??9WJ#^6$.Wu$ͻ [|?gu{*[Y~w^qw) 6Cs~cOL)2nts~gL_4Ѐ@r͜Ui洶BLfkCԓ3f¼wlk˽̘3@ϙ=6:/  ]VRʸ`e=!hFڷ*tSQ~M7Ey9zZZ-+y0j2i=b%ҺVK,{._ ûן.b ï綹(cع(?7+ݍIq%RMY=Mz3ni,w_M%[]h>Zl2 mSza!!=H6y.zDrzȇlK9̗Df9gIpEpzO,jfRNIpy=:wEs縊'6c(dri)`:4:-9^3|X l_3lOy8T{)duX"&R O0$b|>66 G- 7X@d9UQĥ5|&We]LytթM:u NF٧1/̸x_Z^'Q4'<>l3 aMl NM`g>۾ascQ z!/p]6o4znj4ϙ>/2}i `젌w/;>XA[v(Yl۹C:0?38nEbtPUbEt;[عQ%tbIPMy0cOP3`kpV Ngpz#C紏Ai[fNwppvv3e߬ڼv!^C;4۫^H?!rj`N^6!8DN+uIS?>Igӓޤg͋Iᬤ_~ԛv̭/g'ەk MtJvps1Z}xTq`#kwk2-̥U41FzDI޵.J$~vai`L*\U.uؕtũIDQGV";サ-bSUo [5~q1 V+nrj=]lIJnh3Ad(c-+uCOęF:;':;Ғ; z{t^6>t;r&^r@3 Z` : &z8=6}dwъ#o3OB&AE}NAa\HqN4TQe%QQ"]2\(]qT2辇-w'6zq_tza ȲN.~C3|r5춭Cm@ H=m@vrbR4("C'"]#%# ᭹|ekï i;^v\l^a;9ٗ6R$s$oKrΟ݇= U~b+7*Ke 'ka 9K0rvs?jtA}&V]l1 KW ZjNu-( [[v6.-T l'"Hvy?X.#a" #G5!?s..17G1V<'mi{ڒk$c1_j*^χzdpGJKB T;Fvpvx;x;ŏ_:}c|:fj5 !#QVFf%"8KzP+(]Ũ:^j%>[$j q6}%tMDikjc3qMa? iM"tH 'vGb'`Q s7}#;T]&,gl HRF^L<:M ק$Lz{Rang7}YE, o^;'+ۋ7Cy޵޿p/]K^Spg*a]^_TX8^9uK[ne!58xUm**(f6$&}-U0L3j98*o/gecthIg[q!η#?QJ=]nٿ["5{B+Fb'>ˆ~57?nbx ND]6DCBFyUB-gs"KA [u=m \$R9ec¾i4՟6 ;K1kB>.bx-y` XJnͻ˶AZT.h0HJIsi-Qz)%ZANQ.*D=_5HW?lU*`\9[tkpĕb~`T2焚<9 ;UA5BN/|EE-Š#%}qr޺kɣ6!l`(qe$w Ao-KV 25e@reV5F=Ϩ凰M>B.8z[,'*%wO";w!^X9|ՍO ̩w88-9k|`~NKo,q-7]۲~6yZD>0;<"劒*XP&7*`dԒ%GƒZ\;?ZoXJif#A%SyIY҇U(>HpBd Zq :V(?*#~M尘+t0))R ^Z,H]'>DF9CA^z΂4󏕞ښ%{)>yPBT|}[[*IU[HTriV8k{hB*wŗJ5t︓'g)ZI.]Sx%#2Jg⪼O]\T+{QmQ*;%{ҤIپJPa?msX;"ES]Bӈ{2y^8'&Uz49UR|](= pS]DXnYe]ūy.N{7'<&ʝ⸞t?31Z_GO' Oy 6ZX(1z]߇G=* T֐7Yi*}w#hVEV:@SEMI﬊TE!>‰$h)kSFv»CUjL FkI"hMȉ*S/K8 e1<9;X5:шpTw13jިiVF^2q7&U@*qS%@u+3EqCG]SgC!tUM.L$D|:ׄx&h*x ny5t;FR%[cEatV:V$_5.SMԊ 'ָM6 mH1}};cCzRL.{Sb2 Za]Rr1ƜG-Ģւ'Yc\9BI-FRn C^GCG>ӧm eFA -hޚRݤ@Egރʉf] ڪ߆Q͸cJ$תtsFqP$WYɯ[BSX[4 t* ;c溵VvcI ӣ;@ȉ"ŎHPE `;Ǽ?ᱻ!φP4ApF2y자<#DXcj5a8QD͛_b]܁?yMP$_Sd9x8XBSf [B .Ny è& /Dm]P?U#l\e^|ۗVhHtP/>I~y=@:UaL6~l9%V&`Դ:Tg>r y%:y/(Ӳǟs`/F1S`UUcipt-!G;IC(এYYv>o%%[vƐ: |u?O"CV#wP- .Z<+o*7=&wEE<8ZJ~ݑl<8LYypBO*ׅ3UF켟:b9WUOnwZumGI Akld ߎ [ت` BMow *;4x=v^^>T3lVFH7*#޴:v7ƎqvVU*xg>`Oe=/w<+o;{Y= h ,fqoژ:9am<C-$6Nlp&3GcbAC589x/6oD0_n>6`ā{i@rw͛o<2aOy[#'A=䖖E XgnuS>T*w^|Z`j.zkREVןJӯv<%O>}+< S#Z޵mqw& y?IJX=P%RU-%^֥TA`oGeXZxvf<BnX|˖֣n߉z~tWs)`xϩB\r[AޯLČ#5x [提K^ [ABT\yWrsOp0kRJ% %F4{8'fzt}֭ ܓVgfIxR2:Aړ=:r:3}+MN۫ޅm5\jH}h0,G'7 =st+vB1 EF!ǘy nzKDȊq^bw_6EW6-}7͋v_ٴҁ:Ժ /mT驸 <;2ZdE:ẁ:]&A <}o`c%|R'\p`+s50}QR:W?/ =„r鎙.1o[}<aЀ}ΗQHNQ}Ǿڝæ/-n'P1O H!4&3 8!ĸfBh Hπt}c ƶ; ]Nj~@cMp&kÖМp:މy&yiwvi1X\1C U0Y!S6ڈ::jRs{͵V{qig"/YCqiwqRLȯBXf'r )Y4 b"GԨ43Ul6)UC͸T{J!ZtYr K  %+Mik%"XG74(^$s&pԉøMҍ%w[l|=gNiמAAV'WmӺ;Hwvӭӗ%[N'4Շ7};59]R/> ܨ7]0+!.}VsiJg.m1%]W2% zQvy{My?^|/ミpaI88`g7'^'<3>) o}tpco5s8kތ儃4ߺ]Bo;sDHY&,k=i 5;aA:+z 6*Z6*4F̴66b_}krP-'UisBn@ 1k#KAG+*CkXؘm1JW Ek1T޶<>~[gҪC]ώ0*i 'RǠI%֕b6 laKia/5; jw R*8xVsSj7맃;Ẅ́]>;5DHrv V [4& eVhdLgsSؙJөNc%j$Z5hpvJq, 38_ Ȋ60w瘹Jpnˉl0"qK'oǗ\Q:?Y7ǚQkWm݄v_̘5AA:NԻZW5myNPg> S /=Um]dݓJE39pRZ\ *8?% "#N XSUqn>{T&- [$'vzGsN5ſk1 <|%oʵ;gKUxrTߔ/Zp-=TA^[tL3~ 3w9l/B ~~Yw t u!-P)󠸲QιNfW?PfZŜ-[e*AjsI]0SJB[r |jfX@?ݑl[9E+4R3% q8vuB3GnP;%ʢ-*x ȴ76QK^Mɷ:d +q[AQʘ~|ww{.e%-;5{#'p#2urb[ǾWX2 _.y^fPDj쭶0M5;QGnyrl[m $~ 3H5_>%ܬNIn ~A,ڝuICE0"A∶aQ-TGu9kBz[t8,VOٷHXKKfqgߠo>λgBlY)6YҐ*ADSrnuV(M)~-SJgn\Wg>v&&`{N\nW- ȸۯLe[ȞF`2F﹊VC HU3~Ûvvz'wuEeo\ÝۨHrN{;@i1vu;xkdU`]$UE;=뜛 !H<}g[+LgcmVp*9P\3`jh`-r~vC智g!v7L3Z\ywN(inywBBGƘL>BМ9B$ϑ}7Kkg|x#5 +5=s`px+0KwC`iqTEɆRISܱ2$[RXOY$w}.*V3w)R1ҔY`K`-TiED\c6sQ&}N@>b!WTo;LM. Vi}S_b<\Hi=muLN؜/ͧz~7o&WWtO5\=1AuSGFr2e8*癢e+UV"ZS5 Tbdo9(f='470\Kmfc z#}ۆ jNrq`睄!&ᬭmRUz!\˺6qh0[kE#J 8s`=kJ+TPJ81Fg=GFQ<0x \oT=BԌ,sA~\#<q]rio\)̔+^v(J@fs ( 3CjrIjX ȔL~S˷{vr”!-jȳT3r7(A1eJ d_˹&d3odu03`U }[sona9c~/!O@/X\a 3 OݥvٞP؟|xDg[L̀WuZ y6+ fEjMȆg20bb$(f+ԧA6rB^RU|gGs-$B)R"%ڮ`gJ|ƷLG[jd6Us 3 Lo57!8@86J.d[+Dh1JI +Ю]- ;hlWݦ#@MC `ATdW]z 쀍Z@is&)^է&w' kM5%{ک\kxG)!.T-4r'1XExq{Ԩ$GO=6 C^{:OHwoU:kFYVܣ{TX˝SvZh\ݶMQr3{3a +ЮQiJivO}Qoy0CyӜjK@T K2U_~Oggj\mc427d'٧<{{OP"ku:lR om oz5~^wOU wua㽝Xl߃>q/k|^:}a+{&vm آ7m}1olѷ o ޝ{ດd؂:"1 PbÔ"w͐gȗށ#t^nY"%СwŽkW}YXKf;-AR8un ӈy jN{>8Sm0`I`q]H6zÒm\G-ZJx!!E<si.$HZZI,@Fr5|rbY'ph.6p@P6xbe&<1Q*`AC3*^3 u%K~Ў$@"HP @mtئV X)\& r*A}t -N_Ȇ<*0v/r8ѡU  ȁZ{dŐ!A挒HU [oZN@|B bトݟQy)H-&V;O86ViN['u֤^JQH3Xb?M#txnJ$b'`9XČqcbd73MBޭ\ H\/'GQY']V Ȯ X1֒4uP%5\c#`_o- ɠo`MK$*TpuS"16bA-:BÖ>gj-][jCztm'ܵ%: c:Rl0Fkܭt>F(gJNA(^J2M18,ߚPO$Fr '&P1-N bTSQ7Qi0?pcȓ`&lsjCIG sS rdFc}Xr 6I"5=u8$BTVs]%§2)[+Ak:!D(%"Aʣvt%Q{!& L&A2 /Dwlx&% (4<zh` /*aŘԒN4%]bQh>X O+jnQ|蟿>zoa~Eeg#b9s~_ fQ՞}n8} c$VbײLﯯ9LAWaNw|2IzmquSZ]l`(bK<_Sƅ` 1S1kB6ƿhaG@b6>+^mⴴ0M>6H̲(fN5UYzAmH~٫ 'b!%8D2RU`^z\k4\}qF~bVKȕKdT"\*,HH(#͡t p~\Umi"V:ȋϗXLh%d>uh9*) Q4SmAߣ4²iaIZXK ˽h7 8%u,Û7V¯_bXf;IxHނi,"yipw^rQ[VE`BUld@"AiV&g0"AeU4m- Eo̙~}vq409!~= .In)Zf([l `ҦT"hglabr%pp7K<3vRT )r`JU5ZC7շEĀ8LM.s'Ƈ~6px:+r+E`^xx [i/gxJ7('7C'<s]%tyۡe_+MY?'T^^`oJnf0~'y__&Cl [,ۡdۘ̿ͷ/?<=~c{k>^onmY_Vo>n&͖X}WuN&?WZ8l[;\D&E\:99e,so+C% +Khb"}qa 5ٻ[- oϐ#I#%A*L*T{.nٟ(-w'UOC,Z*υ]T?eāӟ +'T)a/)Ia5 c;|MM^4x[\"f&s@j0<ݱL2xJ9 fj]kc|T p9}7%e[Xwމ i_l0k-ڟ&cܒV]v0N-ٟv!p%\IKkf%TK- V~:񡗘^^`_^jejQ6dxy|#lx m)(F#%_л0ۡ W}gxs-~hf ` j i\dXDQ8p 7:WKe"p4 @Q9}ކLٲ5w5MM?zη X㵫{ L*j}nYNy1J#NYB; m(+Gz;J/*|ГqFeSCW%`u'L]% ~(V1wx`Ps;~]X&NeGN]ʹ^|޲⳿Oye'0r@kN[i;w;3xBU)8@>P1&%UȉXjxZ|ko`(ׄIV>m ztnm>-}h~t\x8H+G?:q;WT9,Gx}z'Xm(VZ'P}*8޵m,"{σξ"q{ОMz Bc7JvOҢΒMRDjIMmQyox&giI' Xmq3T#Tx,hgH6* WKjb0v)-%dŢ c< viiyK&Pdճ!4w0JY&4iPۗ+N鰵ڵ#)8J&£Ƈ%NxHn{$9}@s݃ e\֏P7}ɌHNSf/.$haRr8("v^DI2$$~ H k^^E~Phxy ОF>9 (-VvT5Xt\PN)aݐl؁+7~Ƒz|gs8뇍OpD{*< Vg8n Yب0lJX:L3\]aQvB@H<0B |6ШZ0SabQMCb Cʡ>`#U QÊؑ(&0EC r.a B/"^?/Xf ߘnz㷱Ua/~3sbzfZmbywFiv J(?L Qk HfS\%%P M ;)V"wꢮ\U]Fms9;ڿIVpRxJ&Z*FF6W;/|F= j;8{ Ƽ7p?xz<+K?OsJt__Q"y@7b<~z\Fݙ}|qj+g5D鼶{j@K-ˈ滺wa>_:0ƍyGE=r*;VazI]wZkN{IZztBG2ɇ끣cXQ͌J}zM!GZP {O%(B ٫ kzLd/8+6W޹~G!V1AB My9NG\q? 8*~8ʷRA֬j^ܔrW$G`cl_I)Lb*-s:!B)cV2WkE_άix ..c ʖ/cQX\Eeɕ -̇\O2yyZd/Z,,R)q]w.Go}10׋dv=#޽Y>?+r4ʧ̃F f63;O/ƖX>;\~56 2Yq>gf->_h#_aM韗FY!gL~ vIv(1ћ 姱xɾ:Xwܵ4l$,W&\s/6/'}X'JOhP/3.Qlc` \w0׷KR6[v4 )#h1m^AG4l oJ&kK&kR-ZgY%P̉0T@XDѴ$6QD&"j+beF?ռ-Xp<.c3vi3ubo 1bTkP*? hU1f|O|wi&є?Z"(;aûNs{ uJʝ4 /޶cfۻITILTI5Y"9Q$ Xp|P[ AtVicCewԁwfMe5:=\R%ƪԒmv&YNR![Ú ㍋#0%"6dVϳwzw,[g. 2%tkiI v8Q?˳-ѧ[gwty>UˑZjeX&eX&U0eCYg#9Е#ـvKEKՔygfW@qnW=o\K >@e5CDܕMX8Tvցh)]レ`%./凤+hE&JJ{(kt:LYp)} &l $S{$z~%eͷ=8 ;Z;핾)IS&>^EU" 1))68d,<3fi+Ou޴MV[xΨ?s+ZJ1e)h\Mx H|RKas|ٯ*ѽקݸ3G[p4R{-b`Nvn!¬[$5aRuwŗ<#IҺۻ$up DI{w7gh͉~Cy٭mGϬ=Fs1dW_G#F*G};:-{qf>e4 u8k*b6ZZNEn9Y8+O_NmF럺JN0x;U >3OepBTSu0T9%3 i6 3_7>h}w*89ok1tώ!qXa[jU3 Y'T1#Q'WXZ'gR+o DS"T0(d5dM2GbaHE!'|ȇjiڮh|v"M)&-3hS<* QGS/ݭopS(m;bAhEB{76yǹICl3ai-Dι]f34͛Z t1Q!5c呺9 ~4Bz0,TJN4;Bŀ>I"f?8JιH@" |2GI(אS(H+.Hbl>ͥ0>9(g3fh2fZPN9&fS!Ֆ5^KxS-0!`\5MvAy^?Ŧ c>}i0]-)SArFRh1F3>]}{6\>4tyf7X9X9_VW"sce$s8\K];bhuV,\%uL|gkWq"7\[1}#(FXhknf<-OjFLB;I ,"s$ỉNb,tc TcU8spY\gJ"ye ȇ7Xl /^[iNoQɸuP䯊bT_:"xŘD]DkqF*rTLĽ;d+|)|ݚ+ i9Q]׷cH5kM{XAHU^GS}{U0~%8rv icil2F2y]&O˦QSH; l uЁQQx8 !(-+ҁQ Ҏtik!3Y9bm67qLI~E u.=Uv3|D uF(_FNFPɮT2ίS_N !tr9˲=b9'ѲTPAH!NYO1R UCr A.[o!ۍD-PBz(ZFfZ gahNiJ#~*8;R gTj.AQ1 vx eñbF+R^ZbEll3#fFƽr (敶`EcvN'O(hɸ/!~GJ Pxzd\?WVY!ܐ~E(*]SQZCU"@1ߪ6ARUn>~(VCW*;焩3s(HBY*D=7k ZJjE'.  'Y+ kHIzEV?6߉nRġXrs"?|4d^YK%Eɼk2Sƽg ןǣkPPcP7}jn|ܑ{iJtaBMxLJ b_tn̗\<#4\(lSZf^ԩ)ߡc`N??ŗq Ϳeߧ! !%xIJF0;t,_"~ryC}86YWP%}hdio^k,"% r .oU@(/zW*s{O)5g˔xT 3IvP'N(oQxͷo*4Uuo0moAFW"@i@JavfhgiSpW x砉*;]@*ܛ>??`D:ccTsUxk ifSi*uy3{7>3!\\v|>s0[+9t5 lcˬ|~aQG*R+6sSQWYoj/B愭ޅ]38EoT~o h4tFHhAA=}f2%nGd:8BgeLSAБsQ2f&kfdR0n> Ya۶,tl{G!A# Q%I\58)13Uc^ i%Q}7L~2c4=V [Ը7A CNBT^kfED g7,N~] !Ajz{3W}IP|Ji&* PʺzU>N|5=i(K`9]fXKS(;hJuWrr3x˭Wjc\: 5G꣈ . Y0r_~[4cGp%BId 1TbF+D/>QĀF$VQHwNK,cp@EC/><\wM3ʣEy*V:k_FXe;Ih! ]࠷h ^HzD-!Y>1LHFnk]¸BC(/!PI'%fNMS)b6HhԼ93x%!.haepqϡMS**W,Zws=Q;DjhtF@9Q`%۳DzФ{8trPs4=蹠#quk(ApbAt?r.8));Tq4Zy. ]^ozezPx6|v6+O?g%<rB_T׷4"$Ѽn>ffc~W;~oͷ3~;lYDD/VM + y"Iqo<33Į~K|~{ 88G$LcA9(G ?=S:`F6nJ$Q7߃I Rޯ}lNBjJoJ^ %ATy/n̳]{9K2>(3)ǩ;7`Wlih- S 0g^8(F} i_9w?׀F4,γ6xyuoprN3rAzy"SD2>-(·Z)X2k]`fvx])`ʸ/tTi"]M\4܁J0Zj(#6Ž M9MVHŖ\rKN&=\n;Et_*6:e)φ;.Se 2LAF &w*0фR${F-|qAB%#7u*El(;J;[rM9jFT '*$ZZK!ub՚:U q`n{ke"nr7ːp?{!TR3 G,Y~[<%ZMy kTu K)5Sp!\HPdH^5yu)]YL_w9#89t-Xpc>@m-6#B*nF1rLhWbWIU( f3Q}(hB(i|\\+CPge{jŹ6ާDW` %5"X\3 )VFo#Y7j VQPe-fA`ݔ$ϊ_<O#CZ "0 ʸqԐH E%gJ3\:#Pi|$F>kP&xdzts!M#dù G |@Q=*U`B]d>ή`r8$9ӧ*ʼn4J,'Cxg?G& l4xzE&-*(厣lE'PA2z)BLU|Ryf[`k@)ru Ex- 3pw[ARTP)uJ:0Tm-OG+V*ȇ_۳[ ʾ_$I'}d 2oz4';% Ʉ b7j }EC{ԬQQ> bzi\-v&Y5%GdQ|+||/\ eWpJ.qQ]%e{*_SEVnC z:*NZ£W"4_p*"b6G,f#hN'hJJ;ݏo~tR-[ j_7p~Z?+_㥅dnI}c޻[+^oͭA{\ SX9aܠfk\`$I7,m&zMj)EN8ljBgÔ`C(fse68ٟR1qRyZEHrX&ތòrn-|[eYq^,@*X_%;h2Oʗۮ+vд[1+ea5rT7sU~RvU;vw&)WY-Ma~XX?o+iT0V^W?/DD>ӪQykb>i!BU_8Ns#mxFxqV v8eZhO]W\ϰ+<Ly~t;4=|~pnY!i U|rjcŶqc/a)WK|-BUBWaVD`pOM^p""h*_ G4kP`eUI\íwj"rj'x/9ED 9Ewyh_DDm]BkX+I[y-4 $Us@+N)IA QM>cZx>7 49?\)g⊟*/ơwu%$}ٻ޶W}91_Cғ /e骱%UA;KR%KRHQYMQ<33ºd.#޽ݟL׃[GcPPVDi4;vAk SnUk@i N<5, Hh= &ՑF/%V gl:rKf";_8/JPT\زFΚiERr=YٯnBǻ&)U :1a|};fr-% kjbP}6MkS~“*GJS51~E޶> I1nk9!"ǐl] :OF1aj-l`È`kb0^L#VZ$9L/hN !iNEpǯh .N;B҅$ehc .rmݻ& #M]6.fZTC|LYch2t}I EW7qNO:IvN|ݳSJI"O26n`'3.I,OyDc?Li`z'iF3;ӾYz@| <3&,fWϲGξ=%ϯo6|䷴wj8x>yb/|zr9W 6i@vқu3۟.>{Kd>~qoKgד דtHV {C/>+/7ؿ[н@]\>Gu\٧G`'{3eip2%g8aQxVuyt|4{|s)3q&gOpQU h󌒬.7߲U:\{ E.nCjq즚L"XpifVG17s<~*ZUL⹫ 6z@W 6_mzňo"#}ES{`$ u [Y CmuHvF2|i[QZK$[KrK2ǭ%Z%YĒTrKRD&=j=_|{ W(,IWʴFK2\A{qYF9nBSE$a(=?5L͉ U1Us 1'v'rtϗG=MNM\s>A UOG$ưseFm'UWf1Lf+??Vd*gQʼ8@ڐcOHr&|X[zZvx {"clPf tm"0VWD&Yؒ.򲻓H$qcR'19NS[1ҝ)cܱ:)+_rŰmnJ@ .ml>@ @!_`:m*dC p< P{~`dFqH<._aJT[Hg K _M9Y18l؝"X{dFGMK#D|U@EJt>pt̿?V+K5A%/ $rix8!vїE dջ7$i2_}|HT=N6[*e,\e HU RU]bJCeIxO#*`Cl$tA- :eȵ+M-N Ą֜7U;Q^8 ObצnK( =^oKK[t2sgt.e׹}[XH!'.:y;R4,{9ֺ^Ƥ9wivY 9^*cj>?y M*Ǐ?dVԧ6քcܙ$_qВj 69sᄺ5FN:c;YqhJF݌9>H83\{v_dk]g7܃wv=wRv Tdh`O3VPF+xz7H bN}fsem!p3\1EJxO+ X?Zbg9&b˦F9vP)1~؅w3Rtwq;H!8G 1 b;N74VA!d5`}jחwS3N_fdGgVz ߟv~ VfW~n6tO:~G7wpg)I ZMl-4m?=M`} 9U#DV)P9j#V5o΃jX!Ԝycb#konɵ؃䄠׼by1ߑ_tF ZׁY(;՜MgdÁ̭P7?$G| SH N6.򉖚.éG4sMIP& SIkWnwm?%x)UK]>5&HsT(#4 Έw7>=5aQ)mh,sShlNsLcLknEx vߩ"t]} RmIa;ӦTtis[*M-i0>I!p;0BZ$KfzϩI@4{=QR+spQGRNiJSJ֓JZ˯ V|6ZVlϭEBj z iB=?%!E*nxK{EaiÆeRW H.$mb*& PZ:4uW/.1`s̈hCϿt~xt> =/,o4<$ұf,Es F1MxhKCukA`tk)j2MB`OR[yыEyФT Ih( cđ!&TJ:A@Qs&\0ȏTN*;q7O] h q<\'p*o԰(( (4 cJl>NTH#!c%0Ta0,Xt6o]bm+ h ?MBıE&ϸύT(H]]װU*%s;=Yay8n(BF12XHС84490 J}%Q4L2X4ː\,8"!J1%b*Xy`MT$8R)Jae@Xf @86A3*XH`Xa-]8*%$c-TS9C & a',IMFőOLl|XWxL"vvQtT"wNNJ6JÏ/غ\Woǝ\L%~熏9 R Z,V!< i+P,bb4m|8߁yڃ8`…\T;7$R/Bϥ4Ⱥ"B ˜$ZIk/Ar=rMa:?Χ',26:Z8\V0X>T$lHê)1}FH"3D % b$'PӶ=~-}Qc۝M: 1w,gڿ;;?LG_jզӱIt>_>Le0ܥo/yއK)ιԁ^*)##D@P@L 26V0LXiF~N&%`l Wn12eTEW1:t(5x.nouݨ3H~Wvc'UW>βW('e.%WLۙ㧈wU@E6ڈE?߁t+@m:+fl7cOy(faYL8b洑3}p#sJ Mo;Pxm?#(ܲfV!F5exq!|Ҍu s,p{1cY;.aCuƾ4"J؏Ql831 Lc1ToU9.2eeH(դԕ~.)M.|%k5SȐH?D>B؏>{ %00~@*W&БUa ZԤ#c>0ؗp9 #0ZsF I`vϓ#[@dMzMλ Q*&o*Il3܍nWvocrF֘z$J7u;]YT$ٳϱNlT. !h>xZSvt\2D{{p[[|s{0O,:jnh Xϑy.Cf .&5+J(tijWTr7Y0vww.F8QQ1byxe[~}( `FDTT!ؙ1MArBʫ4j#yǏwϣ<"΢ذ|*Lvg}6NSK,6&+f+AZVY败JZvffa!(Ϳݘ)RVjqwmY_!emCX-0p`d t]7 jb[2eÉ>U+H g)q#*ɔrg*#sIE-uB:Qd(HpLuA?)Lb=lV`Z *5U/M*x ٮjMcB('֧LGb1pVrHh$ݻUݚMU`P*0@ֶk7+ Z]&+paWKfl$wZ@ib <1 .FRI$ϧh\E]_>y> *B&_zJ-=b rUJ(yN${c;W?y LalŻI8?E#h_Yp#;^;{ 7-mBjEk.*e2, UAY: #"<ϱd-Z5fitfbnd:[` U6(ie'au̢ ׆^5\Y 1fLi]%QAHN=~`8?l(|} _(%"l; :(ʧw'Z)1.mUL$enb~Sړ;-آߔpVǻNxߌ LqxwI:6nvyjy%(;*Rs=8q<9Vgp*Kŧ?JA:َ#xF0E|9Xj)"TVy);ⲵ'/ՖrgIcio:UR1,ⵧY̜Xۑ,XҶ.X.~ izXNhc3*o_boBAD:;fӝƈ4C?8k ZxNT8Q%j&[;]-zO4RO"oA)ktXL>'uD0-{)TAZG!9A]"`*LP`Net/TRòoJzn^ث+WO-Mꮞp3s-& hɮPz*;@zJu7CF3dX, CՄshM$ܪU!hSkYP0M+KA8IyQ~nvH:<=Gq5m2cR"T*1H7,lޯ*!_k4?t|p3 5Qޗ-fIm[n;eՆӟ3UӲ׍USZ 34N>;+-}쬔';;%2φp0T\l&on"K-!3Pk6kKAJ-rZ r*&}}6I Ф7)W#k‡FkXzo)Xi*TǓNR#)9u5sJE PYeZ*pJ-N+yEРF;H`ƚ7DdG`j[ӝBh:3k+IoL(pOC%EADFYqYœLqvhBBOZqIcV 2xylM}Gq!MQX }3Y-'~/ {sbcŧO)N\ƛt6[tOwH_ @vOO'n/KX@bdpq3kxu1K[v $LaI;fQ̵AcW2R8$eUDr{iL|o1OpIz)]lMv8=FKwYL>u"'yGzGBT֛N4n}j{r`{[.\'D^'%]=i ZAI~j"QgV}Dkwv|k%ma)**JW+ᕚWVͫU&#'ϯ\h<]c+j&/U\ӊshPKq9N x "_XX.C-^˹:DL`^r*$6)=aFK r3BͧB=0/|*Z2 1 gO䋷|/❃٭l/?avz˫g˭Qu[^1m}AN[~T U9H8Eźք(LSw~0L* UsN= L`$" o\z9pb ô%o_u]M3 W?3Vd#[B-g@T2.=\ܮ࿌faBDqsX1$Cʵ{2&ޢPF,O$"X3äȰbQ4pq.=5` ,%Dj ijA>L>w{_57;PK.J@=|t?l&8`6s<4/ቅ+?,_sE ґR=|nEAO?=P}M Ѭo>˟_}?>p7W-Ί.{cIa!c>-1չ\=|(`,+8W1~HR;!ֱ/md-U껇" eEͣ^$=[n Ov^ $(W}}_@;?}xssDX(YE(?זE?}7T|P.݃Ad3g q·aSV%Wy(Rf O]?r[j፟T9oD) hs0]Gz4 7<>DJ@PZ[ri e/`lx` u 5&Bt9Xiv kd U{g9Wb2@n9 kp\&泒ے/nsP02d,m-H i#.qkkqQ|SL rA!L$FQ ndboOĩI / 1ypX@(`ݫʽgD2@v=uH]<_7d5R6c^ƣ̌>sx -"x !f@t% yH|-SҦK4ݾL %9ƌeL @DI3cetf!ITHȆC 0(Gs[$HEg o2\/0V.Pdܷ5~$pRUr (d뙱]Taz$$ۨU' $y&P/˙0/o*#k!u9sJsӆy6-\"~fXQ d6 r_6&$++ߏ!Z7cFIźՎ"1#0Ư/&[ T4dCHqGq3'0d,BL0ݰ787 +H3'\&'ZV*e-VL+&h9iƆeXjC2#\6J!+-M+"Cp#XQRdnEVTE|߫ mx<])yԺ"_-6T18A$nwi>s)Dqlsy4}h6b?<XC# z~W)4tT,n>ٟdqm~ى_]5_g]|(uR!<0?>-J2lPsaAZ`c9ΣSbo].15aqhDD[yGXi`t F/oh+*[[OK[kjd?o}~75ayܮ檝$bHaS*"d4x.#4GjD tVFE e} *Rb-8-UE@j3r ,XV6皓 O5hѝ%ƿ͒nuYnj&XZZb3< @JírNx $91uHErO<5Z+RG#ӸqE~Y0*I0HS+n霻^ Q-XY:ZmndF##q!a; K_x]8%jY"gζW{=+ 01V > $W cKB,y+֔mYĊAIfBǩj l%]dai.we[bu-]P/&#NД ٩!@T0;x]T ?XAP2pY얢L0IqH_ ddY4%qaw%CV"[4:_qf,f=dg,+1kӬUzTnS&v[:5A/bzR ;)2ߖ]jm|I%v\T,ZGvS$bG mS[E,Y(<1@)HsNQC)9f]2n ɣ٤;%/ ^ZTniN5CE-ySV_Fgp AE%[4 $O\DG" EgRYl4>l_E߀4GG"CD:w $ټ.WRmy YW rPYPv."wE6*=RdC}.a4Ѻ ljh$a$/(G{9W 6BJ):-[TZTԲ۽r@[BFT9 hN4xqk}T،1(Lz5%s=|u~_? >Y-[Y|Z{&IPbj& =S{tٹuoOlgȰ1^u?=;;o2*G)Գ,2ӫwAF^(#w>/"*7}5Ńn% (ia &B9^J}n9OsR3׭?K{*}r/^ud__l_3>N| $|_ZZ,,[\'Ϗ$!_){ ??yܚ orLQ1/oGd,z漣=^܌$WJ ;J }4%Y%Q8<MFZԥ~G ,)L09{eU|w 7Ay.򠬇ʃv 8/ *(c*(@ʋgdL}Rցu2,_5%wӖ$ǸJS,IFi+I.I^?=;$)ɔYK>C&RF:}22U1+=h41D(;D>iΤD`2pIIaYs&%C+)ء̮Zl-0m2[.W џk濶1V oo1Q̒2^4}E_J}FNVjiZ4xz}*G<⮐EhpNdpNy:\4$"g΍mG5p$DM ,*Mxn"1 ':W?ȟN^U =;dO "rkMM =ՋuMmFRF$7Ηhή*2$ mv* ^O땜t}l˫oVOi?/VOf^7O7K?MBUbt638MoA{h\r˜v\h4pPTZeM72:+MH|lKS_kP%*ߖ(K.bGE8,z8 KCqP !R 26,cąV F5Py_j'2L0s =ay3@P3@9^՟B9swfy&v ) ۓA8#Qk Pq&B/>9q^%p dЌ}RΰE(=24G{/%ZnB/yFQ#LR_5sA:2< `MDWyޘ.r%=\?h$Cuܛ]|;4pM߹_՛LYsRqs-Q9NcVYV=c"]ōiSWw_+ג1q8lZvX6]GaG1Bh3 r"=&f# a~W{gl&IB=-0 ?|G`3\MtإyK~X 6t;k^ E;`MoלRu3קTsPf_Q `JQ&L0 @92ƨElX1,x|Lv:=vu )-nz 7o+ȏN_{Η__TړE+3$en_o55ƥ\vP|L O8Yuqy?}RxWOs$dRη[Rd'$x&(苖E5#yBnI@9%\o'6MM欇6-Fͷg4j.ZZo&٭On0j]QC#[qPf6cHԏ{I]@ˈI5qf9'%}6#7ϙQtZKh@唓,zh^缣e^=k^\N8yf"srP,;@-|1Zmcf=;A>ZjcTF6w +50r"HW%ۺ=qmU\ԆI2=PCpuMD۲muۖխbf +=z3H7Zs#T¨,Gk6 OҠvi!))y1Q JjZFn;(bd-,f… ~i,%M3`nhP`pTZq# bB3+Wq;͉PüݸJ2zڍx N`vТb;5DԘ'52V*s~ b(F{A/\ F $y-EThxs&zԜ2Ew54>:}T"_z1%q8-BVMFO&oB5V)tѷB hv@%GnG-׆؁Z8ëE#i<ٴ>QOPJ޻t@oSh Ok?˛ {3O?>y~K.5֐'p(|@hv}ik]r_o!{퓞];̺7-ƌ鏏qu>ȶx;5,,[\GyG/\Dd џ$ʹ:jP Gtj(nVڭ\6|"Z%SC-;Pszڭ9SFv;)f6VhvCBp,SJjS8WbBnD4#( eT(5+/8~,+,oɲF2+f+͈W\,bq$vLYd(8Bj)|e˯k\fg=|5J:B'MkOn0PZL,O,#h 5&O-Ԙ=EcV3hRȃh,D0D0p VfұT{sHh}"DtOiVux2,ؗ#^-VPP?ZZldecBvͫҀDk 8DD)"8FPGB֚YtPPsD&$Q(Y+RD@2#a7]B7(W­ MkmH /H~0@.(~@鱸D;3K~̘n`[p~*LH%ȹHcLU3] []. JFi%XyȄRYC86Η.c9ѦZLE.|"#S -`#Gu1r̺yjBH7.>2U1qn[7& RC=yw9FWv=ӥ@Bq,SE,;αJ.oa4>\c n'5ˡ/ ˱}ĻmDB/~- n_" {'xΉŭvtW"QL*EMAVB]s"+JgXV9.DX"EZ k6Σ`jaĊbqsJXꂱ>#@/HiriᴯuH<FU䅹ή05 kZҵΙ~\x߉{2#׉{wftWL{ƣZfY9=, Z.a9qtrwO#Pc D!cQi!aT{.%%ԍFc 2cu}_X;07jkĸC)AެˇB\rVKR@\ %/jC 13H^rJ@w/itpZF+[*i 9F@%CXZy0*rn @\`P,0.C,]65?ZohhB)rb\SrRެa 0ȍ_q* E0Z@c̮x[[=w6'=zcLW?t< !!߸Ȕh4jhe(u1ܵJ̺yjBH7.>2CmF0 5z)sVx-,,a6/x$2e i*-d9}WE)ܮnƮߩXrЀ\)܉Ds!CaV=y:t7<6xY+ڇX; h)Λ/49LiBIA| j>BqHs9Ԁ1XF!6ٛeڲ2m}iNz)̭'R uǒC1~/ h65mmL%%Hꉞ%)oV(^>Ljmr!hj6=(%BDbmx:a~%ٽ$ 80tp.°x:9ʻCDEPeik^|o@Prɡ"㰤u(OJ4m&;3@C~ A""|ꥷ,P.谷^z;4.[zn[8cf_^l=I|h b Lj a%&/0Š T͖ThJ0+KSRdш Z1O_=&n2|vݜ5p5}yRpw65T!Pd5]cM򍋨LQi2{̲y$Sbka7+^'UB!nj?y p8YDG-̄SMZ;u>.*uMsHVk zR<*=ot"2i.F?\YfJoD쀅a`}ҝmHpcO!$nãj nJr"vBj[K󈚂)%1 ; i}kS(/HdzҞmHAI>|IZv@ҸB֗ %aBj m8ľfpHhPղnxެ $lC0P xf0^@H ɇH듐>C$$ B6'/MT0%;PYR=Չ( 5}Feojz4HpRzc;l0X$I:[N 6o$U<3Jݛ="{9Al1{mSӣn:96@%J8xβM~r(dQ+/_<:tU5*aN2❽͏ RwWFu3YIra#5H`pm#a3][3_oq;G_-NDW1)E)E,>@3brUPEq) 2g,1cȹ.B!qApaJӒBõIPG(Ak Kn, Y+lXr,hCR@d{ERhRIMul못oT5u9EkӣqwK~-W|HV3%B(x޾yzĬ~ɥemw>Ξ9oo˳Q9Kyx{kÎg0[5}|?ݸgdOj;VC8շGwf 9 !Lejo޹;Ynt'x.#\ofe,/%)JɊ:oy6p@{]lO;.hֺ=ȏ%cݸ$e{xb;,w{ܓP^Vtf^ *I<ގld.Fcm>'N$w5p k0D2@d ܬ5e 8v5$w]Y}Ƀi_6jf@рĻ*Al1Ճ |VgY]_|c7>}8_|-q}Oz`B:I^ж\b{}m͠XM ^m[:9\qIV1!TZEN|XJzS) BA eO'Wx gF!}ɕr4zA:uŵ/a;4`TDmlX:m0 {}QEK'IQ7v7^oGj<*8[Q]h  R cAB͇6v^=2I9&w,-3(Z ؃ͬ޼ s߀KAJico[y/Y(uznQd&\V-*ꓵ|tHu\ߙ *qM|f;b0&lJkc?N-x_Wu_Wu3_@OǓu KsظIiD~f+m0mhra$ ºjd BG1yz5z[5]%}z5lmѮQc7fxKr׋⬖VuFU<>&Ӈߕ/9Ugv_,D)b93xȗwy>#~=?oyIeZQ"Ezȹ]][?}=&^4Z=?ӹ@z3j:Crz*>y%8|fWg*+&U]~nПSAtb|֮wDt_.p^ҵvQKoFBy@3"YjÍ%_whz.Ѣd>0A':aLQ|S~19lLybٻ&7Wk%"r0 aaen̠q эjii8|**+^N`;UPLS)-KUZIm=/)ϟ@`tu(n-godlUrǷK_E X ,,(1.Tǟ}Kx{w_.x0Ƿ <4nr{E]D+p[u~}|IQW._W3C3}OI%$p2N0%O8#Cp,P^Q[¬1U*~Ler9#[$WN2弡Rq Bc`〥14^w -D"3nje6cY.-EwCS0X#Y?2A1'B<:_PcT#<ƒ 6 CFJQ k=d}QD4;Z0xVC:_ۼxx2ڹa0;`(%tƣt]h\q~dO@)nr2Sʁ}>gD﯃ 7M :!< r 3ꐳȔJ]iwf{ton ~ +{ĽLKI'FE! B#5,pYH&eW|rn1OA[T'SفDqƒE]ܸ};2Cl 3QrabT  'pjT\=U F3oSG{} G+<p2嵐\YK# ALagFO}߻~&h>{ }@I30qα@2`L@#&PNW5KCi GYGŭс9 ʸw"qf2pkxJ(4^ZM %<ᨣMFdiU.DmfKPO&it$ޙv4;Y5CSnaW^RR>s8\mEOv;ՙ+E /)X~ JaĿgjK Z|FNi  dSRTtS+- vQ䞔`T܂ B/ICG)aːׯ|مj0dc+fc"1Og@T'9nqF k +|0pwlލn,qqdf {Cw'z~g 3K@?D 4+(jPC@:gY**HheܮLTmNy/H=oT5 6 7SY 5Z?e$vK|DG:wtM.RotM +fW*f jQX5Q PrQ9 vv,j=݀ /B:4iwM/8:;Ec+maJ veZjgHP ild*w/ڎP?ZX A !iUD}EP$[ 9uјⅇ*f}aYG*o}Ҕ!c2ew؁hllYy#tu qS0onbhރ{@ ]_0F7Xc61n}N~~w ;s;PÈg`x\bLQB!OϬ_G  g[[/31Ÿ7;"S oiɳ7MVKzu% ,Jk()e<r:LN)~r-g?X3Kt+=Otg{U|bfMrX%bD BdR*2A:4,l2jJKStu%ҙ6* l;͎SfsLA Zq ZY^\0utWd813Tdw3U>Yv7xIݼoG!f)Ơ,~^x뉷rʊ[[V^%. Fwv}u).4Iy.QRR'/3./3hOxgD ST~=B}'85Db)|1E)X /filC!8qD ֕.;Pn(F)S"(kĐOJ5 ZWDCD8_r 3F6c‚ų wL#G0S|ӹ@SjH.yCA" FCg6 'K"@.者SO]PA__)fRk-)KK*yUPӫ%יvrKC:pj&hдb.ge¤#S!gNU֩K2w,=VZ'3';t{%<*yDq(LQ 8_彵N}kO== ɶz";j)m{XBQqs@)uۢrgHD2x1fJXL:}{$兝gk\ Y>ecz֏7wcKXD*?V'r~)\2}O%Aw|u]d!& ,TtgnI6[vKĠ$>Ҫ-Uր&f !gdT N3L(:,hUvkB&-vXK`T N3磤$fdA[M4ɦ9mvT_bPtRQ8o :d,hUvkB&ٔFn _Ƽ[*UL'u[Pz2vK*5a!D7l W Ô)4_K!YLЛ_RëEٍn07S~Wy:zGs3nNYubT°E Hq^"Rm陼Ej#%J^R8"v"Y{LDx03x1[{FfÁv o#HD0EI&2 9Y ΤIaX+Gu>C??b;I!L7 i<̀w~80K5Ms._GQe./)ǂ;&t $r"')7O?g,?{=s[:D_k?| /ίo^te4-}~90~hf~G+R]t^x ~~:`/}/]0Cgo{Җ63|C0-Wo; c^H.f`BwE_*%%2HaAiƭVc}8S8I&b 3,ZN~֝K~-Ӌh4/Gnя@߀;ME3QBR\2oI4y-NW?Fn"O }w|O~Ss _ l t)h Cz}iG}E5Uԡam՗:-WVMڣqMm*c-eF a\ FI:2?3.LFٲ:prE}"=*\g`UHκ`RMnmʊemTؖ`xNF2cp]18ms\1 #Pϱ|]j?֨jBJq.v(o 9{51wE<]]3K$(YT⿵y?7 M4-!Z[ܐׯ)9H#$q(qa Y8˘[23"'xDJJdK!hKMpsG+j2B8` 2Q*Ty7AIKn/-QI3E`iu aIц:!-.\1kB"3eAހ;3* E`/JE<=}g&9;W֚EjKORtS*tNU-mxQ1Et֯VLGF/&Uz .`VX7hyذ[+ %@7%w~GW4oϲWD~0^'yAq_In{3_{;qfc6d\Jr^Q*~hnn2G jV,DVY Lc2:뀌gH<^󲱿b8/0fNub1yX-zuG^/}e& {,Rq⹠,''c>nԇZQ*|<䗥m=I^auSp<$( XHFiA`XꋢI4б #&dRD 0 B DO3%&G2+&8_ς#fX/+#hK1OxBe$L pK)ft(l3*'b1LHŘ( T-@߷F>" 0ҜD0f"Z.*&!]2JT` # >F(Ĕ!5f~LVX!u<,(} #¹ᕉ#n'Y: _oPݷc4-"etd=xd7"[ uY8VX/w׋"h_\!uŢ-R 1vG`C)MRy4G9DQ-EQ9"jlKUk18i Di\|Wa]3tnKGL_2*vozqq(ii#%Xk'jx3!䰤7P_*ePX C-XOHHA a bN/<}d]Fb¸> @{OEթ7)Iܤ=awvpDs7"ΞCQ4>Df/NP\s8"weMԉ*@FE9})"SxCf~h̉o,1q1vcW{Þx{I|7M w$%/$ id-J`K$ qRnFw/'-)BH1w’|rҝ,D1fKYUA^?WRf/ك7kD9aa1'7*WxE_Ϻ֘!-.,1'BTsGʖ }OgzU,Xs9,͚(TwĈ /.0QH -W .LrgV4r!!n6#Y>9 -9µݜѽ2_HLBlteQ;yA"1vEjcX`SƲ1 dpyhS4.K@Hη4(02o6A*WBR 98ʉџu۹,킛t'z"W֚fÍUT`G=l;--)B #⾠F}nflA!tte,ny=]q^O|a/yj@" s[-lV0ig#zVM:6u>ʅYS vX/2h)yv1@A|ܷ9jZi) S`)X^ӥ"Vҡ)h=kYjK /΀L(2YԿ"AL9޹W+1A_Kw6K8+>27cЋ*RL } FT 0Nfe\,O^KHpLs z1vi5?_uט ДP@H#u0CtA{&/nvV+vgX%$WOxuqZSG`{1B1k\G~<\.%2Z iLs=:c~_.G(=u~V @qIoxG޽ݼa\}GA &V\lNݎNm5MLl*54k;r>x1kW?EzbQ~ũ:.zR3x 1]V.[T2(v&5^@Z<6[>\`ŝO{ZauS:-LÖ$pp>3Nzh6p)vU Y@ׂ>NV(p=|旃7J\dRʯ$3 \Y"aA49G ?1"9{1sa#Ϧ 'g1GzL8dë3$2zWkl#=rkYA+;_޸D2]{ڲ&Ag T?)U3gOvP<:fL/&0|6V$6oU .hΘzj+UILV"٬8feGƂ9 .l}7]t<5fTZG|P0-92/[>ċ."(MA~,"-%%uePZguA'_:ICpi͹#zU[-Pe#|x1bT  B8:gH-;I;j!g! ٔoT7AC."BΩ{ VKm^Xq9|^}liGsʴ AKUJ?w)ao- $CbjtMpzNoaLnܥ2:ZL:D`H1v=.1uϋY*헋C";G1Yѽ{!iI|`q@b9*h':YHIZ*\< ^(uz ?t 2\Owh1D x՟ˇ#1)"=.'HY筶B-;R}ˢzU=\LBlX e3pfW6<+-s5j)W|\$x{Y‡5of~v+@@ ůRbP@8W.>% YJ:lS[R9L:',zx1sR8bz6'Պcfڃ7 DqHqP#DF _߂ s~:@Ŋ!_c.qT~͒].B+"L^VYɀbly.7'P9Im3-z~X~q5r9KsݞV 2qr3ҿ%tRxdpUb0у: o4<(v$Ē2j8@F^S_Yzav&ќ j|=d?!Z\0WOn669|̟e L<2&aX)C~8SZzlp~QE"_7eyMtwNa?y\/7N޼#?5d[OneR pwkfb7߄R]ΦYF-G|&gF09ٲh*0Gy\Gt-7E!4 p*Iq>W_:RW 9]J כI8kP|? +"U=~e|2: V/q6WL'/oALIr6-x&/4yLݑǣs1}\Fܔo+`156j*c2xf 0s5|J{|n4s4?50|UankʇsxmJNvj-o Բo'](D k2')?Wt G7sCWb@Ow!)qD^/BRN5g-`KδRn/ GHu%lFF)Mew8Ni~s}P"_֑5(dv>y%~k֝ٺc+!0*Oq1Vy|\LK"14_I_3scD 9~@5|m.\[~0+D$I9P|m@Jc͇:h-cS8sqѽզ[J;R!%mI-!T2g=*\Γ" C&ꑚ)L!HNHsUkRKNH ϷVrRl=kNA0,ǟ&Sѩ*Z5(Nz6?Oec_jrߎoZyxY7W 5Mnh *$&p#a2"5Q8 $SED@qTRT=7NS{KOZMi7:vF4k$@}PTڞr!5WRo&B su"'r:_-BKBid <1T8{nM=jr "%*iw+S)o֘6ntqf&FA4$aϻxNF +l8f GWd#M`iL%Tڥur6pZ.J_ $aѸ;A xDXciYBRa>.PN6r{_>FC[GM4["2p#Q bʔ3EX5;4.X7TDWIÆ6+)c? LQuk'J,j銄(9QB;(8ѱ!pfYa~̾Q:,8BVѶvC{B3ԫUĝl/dJ9^l^Vޕ5q# ǸظÄٵx(%f[$[*ţT/=ԍ @&)7(|`sںA1BXj v&ȡG\dj<."1Fv?kN>ci,J;S;?? R6j}wϏ;=`xa9 "Z{?k.V]$GZRSs<Ɋ^aZ2O,Evfb,Fma>d & X:F*,?dېx[ptSi{ݢ1|妅yj6Wl%f}>,_?8 SXR JzSt>_YdC&D0Z\cqtTQ4}(~MJch0M| ,h ϳSUЕTmeHJܞjy㺒JC%=ֹ,+/Qބyof/YzEt^QFe#O<>;^Jzk?>z[ap$r<=?Q'_f#F}Q撸ɼCk}EƇ5d-t}!Z{ѫv ` ?1HJ](Ͽ\;!K/t3N~A3ؙ͊bѯ:͸4XS$#k%r`8J%g,UKA t9 ZJr.YX >ҰbCw{ӆ2 9 @Ȼ߅n97ᶧ lߥ9%.p5z䞊܆?[G)P[7]}Aw?M1n""A4Nb gOo`RժSu䙀 ;f!{_Hg&DMbϼ\$NRJ2{X Y::as|:`- fw; }$#BGRqXG ˨mg( +څ+h} MZ_0&œwI1|@5HkI3Ǿ+|EbSof0a>?s5pt8seޮ>[l2'77M"I;XФlm.܃ͥ'GmD@K%_ofҫc/DzP}/]{%uݨԶ8)v0a " a#8bL~;?,<`YC@+UHOVdZNRAXjxȆLU ´FGa R%RkjM">{ǡ3-(e}70D +r f*tօ+Ŏ^9!Udxl#'eu4y:YꝛuXO#WmòpFYx?[sUƍ"M,9INtj a'0Cg PuFmVTLiA I3}H51L#2U\]0)%ҟRZKy2w_f~Oւ8)qkTqPk7CTfe 1RB3A\L:FFXfGh="HH}3#Vګ~N M{ w>531 hR5U)}-a]]1wkF4+҉؃6_w)DaQC4~S[7!+68Γ&ߴ$7X  $5ٌdXJNNha-c̰V3;K2ߐU)8Lsq4Y>NVb_m3Y:ŬG6.GƺP/ڄ<5i/]z Loፒ䩨B^=!\QA |Ա[NLDTнfgT!S"ؾP9*cFu xN3ysJh.VeR$lQ6?Svjن/3yyظs5C}+Ao7A%y3nqvxUR5 ~l; RZއ{RvD[|^ uTۚ[\3ñƴÞom {sٗ.˸w O2U(5'ϡt~Jwq%ũ^z۰PO]tbV|ejos(-wWD]qʂ4u+%??πzéOP!)PΫv'\A+EG{Pٓrqת:XU0}6kۙ}rUt_&hΣP钾\O*RCkPP`:VՉCO]x+Y "U zzХ$WS"39c8G# c$Bexꏱ٘~2ʘ3eJ%%.|QIΘZeNEڦJ ȓ &* ngε0$ņrCF)J%(i9q^gDv8s֓6VWӚ>uJr:C!=uZ uo NA3GLO<ZR ֒*0 +C{Ejn]; s39U>˜¸1TVշ/zE%?n05 Y<}oq\`BPmKtm[>)1߫-gX ԡ$#ϗ(w,8iF.1F9F3ppRutJ)sæ J x{񫎚U7\B:PY6I0cy}M cl2&jaY ?CP*  i*cL )G:aM(izqQ$papFL%JkxDkʔ6brOLX4s ֣Y{_)"yH!E`/ܳb-g!O%t}m7n/69f ᰲ 1+ʎ[WXQc 2:ʮQsA'Kl ƅNaηt"QWܵaz۩dE8+X7?ߞ-ʰL66P_d9fE-z%}A-MRYIL 6g]f7;^b)!9 G#)J#ce3- -ejQ4t`bg ړ4,D<6f`&I%3I3=lf Yb8N8`?ػf) v vC& 7LSRa5JH9,H*jL\c&Q -,Kw֫o^5'*,5zg$`XܻfGI#oc5Q9T$XE<5II3S!MY0T&%5*̉A! { :+n|.L1eRס(QuX*Ƥq9ukEޫ+zp[B7jJ{}_cԞ ّ`*۬6T"eaq2I˨˼u /IS N6:0؁< V}.E! ݭR)abmH< zQvr>6N"{5M,f(aRD b BHNmwև1-^|pkS}ę?>9P5JxJgMJkce/Pn< ?%2rMݞ9ܹݩˉqCZVӇ :0Gɥ+_o49U6!2/0_X'zÆyX/@"okd&2j늅ϭ޿"<` en#=]ahOsWw![ӡc{HįSLހl0KVZs B|klӁS ý^t:+d9@L*L1(MeD^FOZ'i-0hKd*ܟ5[4Ӈϖ ,f)x>Mg;bUw*\UгsE<#xjhIpeY責bZzU8g(c{+]Q^CgEyNDž:V0/}6? Gdc ɀm|bq2ΠQ newUr9JKuaaIK ܞ_s.#[E/@ E?i2ACf'SIza[ȠI sXH^A)z'簐) AB~[i4^"Us^ţi}BgH>BP{ GDnza7-aJFIPf*4j0/#IS-ph1051}}F3orx en-ĶU@,}i~ 1c4$4+Aq?(G?odOpf'Y>/ZTS5omո)_}qrJ)b\!]Cw_̶JZ"ว xgmEg+2 TjV9KB{‚S2xf-4ӈhzeZ :Y`7_ܴ*{jC -R @'+`֝y{8+mFE˝ C ȠtyI`puk-o%>x#/Dcm$:р|ň"f/- j|b4O^StAs7&>,晌&4 iNόc* 퓫zݦOHwuli~Im}0f 7OYLV'#-0y\{c)\2({[%A51Ļ01nd?ЊuZZf~b!nMcQ|\ʼnpTEZx*eT7S$?7*,oC2N.l OT).ٖr7fȇo_Ogj@|Y:m8F0:I4ɓ7{ 29l8~JxG:r|1o|w9mNz1> {=k]Lֶ9Y2es+ː(WR}Qv5޷wBZ)yņvZ÷ )֬keB9R\8 w1y"<&g.)T@ WĔ9)OE> )Ĥkn g+Λ ) 6n++wt)6+sj拦qrvyJ$ 2KCy!gNX˔ +3AŪ}ZPlZBc-q0k.3<>1q5 ѻ7A<(71|̇*oT@no Ҭ)_j*2y=XVPԏW^B<Щ oוƣ9S?~/uJs& QNptgL  J:ؑK*֬;Q k*ys۪ж c-G Y{f-!Z)7n~6+:`U8E.Rjl||Km*6Mwgf\,DpV:k1- hAZ4&Ymoi(4ǔLNCX2-[p+n^ѭA]u܊)̭?&ҭ!Ujutí3%TɃVv3J ggIႦjB[:@3h m4U|Qm8kT4!/2X,Z݉#BКqPΠc(6Һq2nY}Ht>}&1jP,%"uc8X$D.c'6o_7ii@3h Q{W-&znҍ;`tm?K6*Qe; >}"u`O$cO{"4\C QT?8K"wt$t$SrgbV}F։M1O 84TVKr#);K(QMRo)~hy&ws);rBvHMߡ \lֆAp*H(I|eօ&u~YI)5;˜lٙ/j|{;T! a+(PdW,BNh3M8:;=3"x?`cדINi>C\B {~{x:2U>*q?:{N'q4~%<7HKp3l)[ŭ1ϊLV< Kd#,Ko2}ցJEy/ ޸Xխ4^4Ud8~0L7$A%# 筲sE3)Jj$ N)hţ ȁf{o>h?ULᱱH]&@ɦYߤko[EvhA-GP12. hvf(ȆYC5}*Vj8ݾYVbY>>(nK0}'·wZ g3(bOmr+UQde+AxS3@Y/YW6dwg9Z"лvarWo?KT7}_lz2v뢏NQ^Qxu f0lܜ)Iyw=&ZS>x_L0@TOFyb/ ϡj% ^vY]gҾak02Jd?`[ K݊U"o6coº_\2)xGQnqъg<qpϨVIXDK$Bg3 yN Hٔ.:9KBOUͽ7dg%o|{R4 #xP9荈Q! > X)x(VžڽcZ+vp3,Τy4`[$SNçR-9kV@ڪFoLh+W#{ןwWW/CG?NM͞`LWmb h|1;Q]NfzH9ozp+)پU`78voA0F&0É}=K3*d-n@Xd8M)7;s.Ha inRًLǝ.Vu{:(Z^j0}_j:,[_el}6ZbEId2q0ŐP+6`$x:P 1RK탪.>|c4kJ?g4_)âh=,hh SsJZgS)٬ʽ;'DXAp!j~PG6Y͵Bs![xkorMp@fRƘ%n)fCf^r|*KNlQR.|rz,Xڼ:ϸzAN5}k\yx7][kH`_WݪԝFŴ|W:O@ u'jsjCB0K^QlLkM Ii,7EPjR`+; :Z٢Z 8nèQ57?Wأ^+ajF/@1ia&t "Rf#-BiGu$5O.ɔ> ygّ#T2f>`G5@3m\uLh7 9d"B;,d03B=}R/$kdmy2|p.K΋}TJ[HKL^#H<d:J\LU}flYk0Wzy0B k aG^r^vɘ5LI떰,ޗ[WVuV!u!}{=n`V=|O!⎘K5XQGBiE X]ʼAPIfꇩd gPFL=gKG%-fٻVn#WXeU;<_T9Vj+^'/NT FbL I=!)7 g0$u&ΑD@7Q8\T0b!VM[z9ǖIC3#V g#%\+[b*^ ,+GhFVRLYWƇ!S`É!+ O8 $xe`21` ~sN%oJN>ağDaUz9[[xISDXeDaޥ\ÞVmPvaD!އ +Z0\ t;(+/JvY{MuXt@'* vϏo}qVb97*bjML3qWŝ\$iR|ILER"XƐSUtLn޾͑!̖QLL[!m[!l|Ÿ5N=8eVdHW,bh帟@CE*(ZTPREM$ X#O@`wVpnu%O1ٞu{ $qƃ>͠ߏU7Vu&CؾgKf9c<_s>͡/:Lނ&3'ǖY\%/Kњ"k !=9%\J3lG=QF9 x6bӝøs *ʏ+GBӹ:ǰ2!^6xk"Mka/9I93J”lC3~3#wg s 3x 7٨m5;"ӝûs(<6¹㣝cC@0U X_I3z/s_D>y_EIԘ-翰!Micj)*s4k̜/#$kR51فCבs\DejΜ+*[)@VfTWk'^<T Vj9vݧ+sE/e*KƪxMLONn.g~qNKc榚D_4cA_Ѝcumhxoho;#[rJ90ݷچMl+;(}J#}T͜D#7V$O2d4T:ʝͤFÝˉ}wʹ O=5 {i̗$%]z?3Y%B~ݒ:됼F`yZi;lM?Ǐ B"TƩ5b["%eÃV U,C"aT s\"H[ A1WI\-2!se:+}CŎPtRߣzj fk PTYt#rV!029aJSp.,T6paWroN=-pX vQh(Tv P"; I1*y`G̅"_KGɈw^ZʂcrX,&(9e6g/ea,|!!E?ooР%Vl<$}a5C?v[fl9&I*dnV1@Q琢҉#k5H"#$,8:_O:7{^$o>tR ,jB(k`(JZ-"8(DѲ;Tc]kGT\NT^fT#$"at{N:J86 v|qI$[<2.1x3(|PB}l]crfP@CFE+79˙̒tëC5.?Hqaq;*_>LǯAt$v82ĎGq1gKZT+@zhVVvW΢,@XH|,sm>} O $$HfhgCnCf<5s J!{?$>Sbp³!1EN3q^}R%UG5\JHp;` 4C귻`(͈V \k8zk Ln4WK;]<4{!NJS wD)=|7'"p1brNE`.}#PZI[/^joR[PB i%LI"A]f±4*gH2r! c68CD2N~X}QFȟfLE ɖiͭlO5+Q:ii%9v9K8ũv()CYJs')إ83*5(74yeԩ)s &@@,KqMZ@,J$1]yJ3L8; HpL) TcZ# 7P"T͉CkmW28 Hh8`8NDgL"=ZUԚcކU)XPj1k? ܠ' _%UWR D`EإRr}.ʃj,F!J>b։6w'AKon"u v^.IҦO^\_\+˕HKe°"9KR%RaRḴB04+21㟹x,ejd@]E@R^5q>52N9^6)Dp$LXؽ&a0;+FQ( -}}IA4ʖm˘}͛!fc_@IIb̩FGlod4X'&.a]ScF ʟK$!$4Zn HpB,%V~ A=nں tv_t#/eg+FH/Oy:[vC^uOf wO٨^&^$F1@ d\zqԲ5ꛢ;}ΧJthXdh)ѡbզa-[+c﷠{֨(V>{<wn FL7+Fkqy%YNi[y锥 |䏐! SE?E $P{;,_@i\|UG)L7bjzȞڇo[LָT>J&gqI@P!/'5+J%} !̈́D/c0P'z´mУJB  NPi>Jn&:Jߌ΀1?Gs4>NPD#ŐŻ?x H.WO12dl+(xѥ{"](9t>mDRswt3C3RtNr 嚒sp;1Ϛ(?mBfSMZIIRƮ|2 !B d%kcU\AI!LDI:|.! |Z@baÚ~߱ɮ3B;|{d<§\3Г$pО9!;t-RJHKcU8qik8/bASҡvg d8x)ry_Y,aб:13!f%C^C8Tw QG 'H[Iq2@fj{_uY$df:V3H&-mIVKlYNOK=dUz MY?k|zU\ȾsG`,owm=0Z#QfDO{eMxVMRƝ`CqBC49FD_.t{)mzjH~\0^Rxq+d]Nəmh1Hˁ/<-k7\&?K4J?5Lqhe6's&ˌ6NMdڸʄ4s⇴$ew\}|s/>LLn'>&/6</Pn߯Yѵx9&篫=xnWO\ E462\]pcbH*4g}eN<Ë5Yc]]lmqKB!9 V EF6:O4{ML71mzMㄡ0XL Ҧ`MѲ@NA>y<Q<Q'ӵ12D^̰/ǰ Iiî&|y S[K<_.-H8[ZF[IfCi2kl--h`cb);O!ݦV˯I9iA~ڀR|@:t>E`3nルtkz1=M aJvk;; ^y~VA8q~a|^m9X']&uYC̘8]NkO;-4w'`]&I^$n'e=VQXj|7On y<#FMO"sFϔv[mkJו\ua'd_'l }7q.؆|ߖOݾ?߯l\uComw>ߡ_;> EK%>=c);N1"Ob'j㱯o<5\AtcŸ%ۉ0 P0GX,HS&mbcAM2ёcIxTG48G0D'Eegb#“5Xxy'@7yyOV2mn<9uJ敚SaW괹RYdxPILK*5N[Aˣr>Ԙ++?ᤐa v)˰k/Vl%RìL%<ëkU^,ե2Z6.ʺexYޤo:YnaC#2NF{)=41 Zp5vZxaRu"952Jΰ:V/Rvɭ឴i{RpeƜϟͯ; }=PЭ_^FFw6zcn-hT!͗=m?D8hq]H]CC=}sf{gF!% W #I}hyFH凇/|=t-6f"Y7? iD&XLrY D_{ǸbtuQD9-THlCo巷1]N"hp3wg,]Chja\rZfO$X2:RT9' +fkQ¼3y DS5FCkXPS'c R,D *R;g5HvӺd&)=-[iTJM{Bm0FV'yjU/s{21^3t`yp&2iP[ YQH'{ań^tccA2qR!1=VTRbkZTjYy3fSz^ hbXZE&s!hSL;U+%}4h蓮s}L7>3OOYW/?>y%.p+~{_ѯ R2Ѐ}&3VR&_?=t'aJB!3XQoҜT\ͩK@A_(.&2TֶHR|^*Fc$#/.ll@ ǟ#f\r5 Ǔ9@Ca1^[ o/3T=*Ca4^HF} F1x(Q.tp+P7]pwQP06(@amAO̯e.3k_qkd'TV+AO!D@=STÂFX-$-J!1A'nkS1_:D{ut[e2pBK4,Po#A[/7%Wz2+5PZZֲ/lI2bTB3ͤD <+>ѺǷ+=UՏGLNSȢQVӆ|E![9АʖH ;ui_-8[za;쵾O-c ,w\N"cx4=l"MC beA NǿͯWy ۽An{gv5} YU}Կj{zwHrNY뷻_>$ }Ci@hQ;|t5{KbR_y .5:`SB U h;v4:ݱK A!RlwB3;# FuIDzJ?\B)pD8Y Gi;cdlOGwxDTYv8gm22hg_Z<\'ZVןPwxjg?+YȾFAg9%So\,ﺌr˼8-(&?ݾZښocPYm`*NT,F*壨vJgPq(Zz$f9^T6LzR A[E:i:Z㵶F%}MB#:m0&Ag/};HP3!͙cnKZIT' 䍅(hE6!CTbHqo -0^8hb\Ng1YTnHˮbGykaSR4VKbSp΂or^/wOtvu9WsB<:M?Jc_psmgvis2XΊs&X"]F ߤS.֔s!yѣ*tVKgTE WXI \ Bё|qpǻ6;磓䧣׏V`,0x~Gϝs@g sjkhut()88^>w=jOg⃫,d.zxIڎ8[1嶳sL`\Mog v'upʟ%@Cڦu@OD%)݁ƨE9&gY0t߸wl,SAIH8~sY`IQx@9!LJ=Pd R̍c j]ҥJvA>ጕ%5#fu!4Vy s4XXI MR |- kX7[Ë,Uâ L#S$z+G9[vdU WONh >L||BƗ "[@̎3^` z/0F·23FF)SNCVZ !GBy_Y4eӍq,_T~x3]\4Jm Ʒeѽl50ņ1( b^\4GnyxJHͻ$yGuA`@awl)r0^zB Skߎ6v ,[)y9Ȇ;(P}Ĵ+.HyQ6eAsc>W,rGE'1Hk;ZAn;Dhwգ( П iYҚ䙐yoRXuj;kx3 PWT Nivkʬ=|^,u[UgXiC,ͥ7 7,Tu~xWI:ՏVڭu#RӠ5g&g<}L8icb%3QrYuKȄqۘXXfh/~7k.2|xlKKM(#ǸgmD$b uG}/iOz)@Sy$$`lDN%g5B'JjRxJ)DI7hRsS$c;R2HȘi%7L`9zN>Rs6& . Q9RN 'Cb:A'Y*eCKG;=qdviER*ECf;BV3E:ɒ,-"KwZ'uEBI1ˏCevl0IK۪|R1^QH6`]l%ڔ<{6c'B@O+>]"B ֋Q km#GEdދ E,,&/lG''~-nIn_U_*Hѣ @.IXdFm&0QcgY^ MeeTg*>SATŝe͔mFVAȔ240Fhqg٬L(~$DzӮyƶq3R6bhܗPrp$$8ɠ6OG1Eg+*$s [~P;;S L"j QH uH`p*GITdO2Z3-R AI16o=.aFs B0Lk/L~yBjKZ%|.bE+0xDd #8qZtbBamx*c=b"B{E^4%Ƀ *L`^:!\np:R583_>=׶zDї zہf|o 902c5e5k39BnrM=E_3V5)- JDN\!׬)BU!Lr0V\s5G*lީ$' 30Bs>H9#($*i#EE.EXRrϔ i+` t$h*3gIzqYby'm8KS(}ň(R%G㑀KGH yBR8|Ș`@NE73xFe@p"x/`O6I+'ȽвFg5@ɩ&LkuFzm;I©_#/֤8H!uF$o& ] ŇZZה,1d4bT1O"[>NhchD|W, .CEBt&iN A˼/8^1)$+#s6gAmZpN\T՚`hR0"%(J9g Yv$\"ZgB {qAG4WnNv 8݃IP|wO6k1́ JRfHgGFECU)GȈ`3Dgi"HAI$Q .MλSSH 3MFQSPmqة杘AB|ӕR ZL78`DJR#6 QƦF) " o:6lC  t{Ķ靐79|™Cm :^=h:/&InLhܫ,hU|͕EU«BT q\rg\ #\  " 9M|>9ΰzF,z }>!َžvXGͅaI5!8t }wsW9|` >5y$I&LdBL t!lsWi.?|Pgj*3Jۑ m;9ZuöʧA4厘n޽k(>' -7WQXa?$#=w|x!\MQ^M1z$urhv@+`Kc%je_󍫫EB97B\k9 tVU1 >/C]7k!sjuσ3"׷@2uz.zbt}Ca2S"='1Nh,p?SIŏEJ>X+PwN)wf r4zܡs_|zPKBufm[' AW+U"hSFG 疆d' 惃?2=櫘|_-M0YwswϷնbq8_~Ow/5Tc+l %e7旯}tŎA6ZWW{OvRw*ѣ+AQ$[;/.W$'f^[m矬% uji?ěwY ~姈,' :{z6RA J{0Z܎p9tVOKH\!c=u-_ӷf4sGL6E_,y|6~"fB *O7Y; 0p?UckN,ߣ7:n0M[ 0e* ٘\k_/9Ϯ#_,YdۻOU9jQM$Ua"wͼd`dwʽ靹QxCjӢ'r;[,6)fAuť'Ni!ֵϯE{ʊ٘ZJkў^%䰂aO(u: uj[j G?ҳݝ*ӺND$ eTnd^i$7zK,M[RWeUaTת\bIOn](2[c[23opO#«CbSgSn=y2Uь $8\)`>Xj%=em&wNL>]-6Da^%''JPeIf wM{ YO̻z3pt>sn*_?ll8j&N0/$19Mܞr\|Kcp u>^ooF5RπɈ?Sq2kkQsՑWORH9M%|W%ꈮC"|*Ȑ#.4"%K|jm,+1khZ)s2B^)W_0z۰*]jok^=Wo^  у`dBԵUҤWEQd(E-]k)2jTAPb % 5aÕu=d$݂"2nL.r96QA1s K t s0A)Ag!h5*!x!l1Y(eh\է$DZ8Cre|Z|r1޺_>.6Xd\U1qY6s(<,*H~Ƿ\.%#(x{ǵcn1C'Ov~x2>N`ձ_+ Bqzf%Z ?ߑ&HJ̀NxrQʠKS6^L}wl^7nT+Sb&B҄'y;F oO$i5=!QMU_oRh&ϔh u =t ʧ~Mʃ6\ZIAzVl"M+PC_֌D)-QȿR @NLaOq4GoM mI V| I0`n?(J}lwl$%9-pt<*VX J`_x#V Q3\;3IU;l;l&Ms \Z.t. o_7h]izpgj/C0eZNPB[Mr3͏XѧZ/SfKw[5p]WfF\Dϗ))3P }^)?_sΑUzǭXV.ҟ*b2_.n7/m.&0Wi~ֻrv)هm#)2[O^z4A Ȟ+M;p]|(_.Ia9#L16`8MX{%4 ebItх&ozN^M0X 1Ӏ>2S 0RTP6+H*%%L?t;|'h9('M^hMJ c M a&gX)ԹZ0Vڂ[e܂ˁ7l ^(ZLY# "FVkk؆AOC(ުMo] zneK]þaDI}m(.6+n)%k~a'Dz]7twi<zr3_˖dU۞.HB9 '^KYF}ڵh1`(GνBLȧ_0D3BR z~\' D7Kq4Vwb#9km*#5,loo@}w{fzO),-۫'tMo z[o3'3W]Plpjr(q{*Z[,C%g*{fts:J!@6'XnGq[׏utۘt) d T\kV SVi BiT-e!=Ju+\fF1I=L@J4yH13P ?2+c |J$@Rx9V)Y~:[fEAF\f XtU9@8+>?p"ש-Vdj}M8nH>ЙH( ;%`Wk=0'(`hhn(@#)J(VBP՚W=Gs0Ӹb^9Ą(a! G\pR@)Ÿx 9+ALRLp{/4炥Xl4psk&@PE .[{\89E9[p*9<&l @0y]ب'EzI_ `P*r P8s .r}kvQ Xi3ˤ 0 H0apK6:Xa**‘ϕΜ{vUo-MA) H'#5`FX N8k3BĄ_ Mb/r V,d!qb1$PcsF1y&P!sw Nﲻ'T}j M29fntdHM9w*)vk΍PȭxD!_5N%AT,%cR MW9tutEY =Q̮'2Yzu;?6 aYNj:ٶ?\}h国t0pgCyB~Q?@_!6ؙu#LFҦl^p&R;`AEҾ F ?|X ?h}} KgB.ظl8G LaK`t a ~5 ?/`[\ϳWט8 .n3S XGDR^9ZΞ!Hsȱq.>=ہ= ʁbTTٹe ]?,B`H N5}Q-ܯ萏gpAH5;ynj7`;%P!!-}JK,̵^*ßj~ѿ˿2B0RUNH?:tPGZG 8@NS_vdS/$ 0"-},rզڐ#?@DK.PC[5 ¥= ¾` `-RqX%<ޜ#VIA퇈֡z6$IC*WɾɉTlbjb5B@L܋ʂ]SOݫōYOe}~RWe[\ɝ |Ζ9[ytyrY+>Yl>ECBVzղٴ/UMkBZ{=̦=lqjA*s]h=ӮMk?V ӵ['`g@kѿ zxZ/JyA?kgd=>90Qxn5R@T>J;y!+EqRxXto@!ÌA)U\Nige$AK2{Dbs^T|51X:5޸DuHg06ֵSѬ݋8i@&/6dF8{GEFrXQDgp zutyVsٮ#k n^AGčX*cjjn<~{1 ۱>Ŗ9w0d-Gi[W`; tGؼzK>)}ơ v!`?9}Q0|M3wg6^<6}wp~zO6Զ`#iG:^ 5wMBbDl`b5rz<4~$;sPQF\g[g=<2ey]صҷgg2(f>ΖM~wt?t  :/) 2|t"U>VR0`8=%$`X:JI 0vʞJa yxR("OB!i(ݤ vX1*r3"襉d="?Dh9tQuƙh/B ^$'vY%9X";>x)}ڒr:\RƜc384^a*Ybi]Qc^sR<9.룯 )}%`4zgQ*2FosxGF{<=qtQ)[@ggdQ?w%oJL>VIn>7Gd=߆&j2_&t@Qwvʛ~__ۛ}ӄowWo8[XtYEBM fy 9E*J0-c+>9v"9ޗu )ӖH|E&uHʚaCq#xjl#VG$d"Dyb:!BC7si?GizLEzsI2@v/AZbL\wFQP' Zmr0,:K{t W>lhRUi U8X`NE)L DS#,9V#$Y s9 o,0hh`xlSljWle/к5N\FB>tS{+pwrOoN+Es"pʼ#XqE17&Y(P_ ˎ1LȏγOe#1O~S~>,qxLp@z0\~8r9ړ[/Jsg\Tw'}$:Z,`uRaƂ且+cʈh)wTTc0'T򂔀/Cp\%&p 9 HȠz08B@u3HkO of`X:Zbe%$ʡx7oP}LAIRiv3j0͆2LҜA4˔hrIʐ}9´SͥƳ8R6[oy70h/cAybtq?w^{:۝NZi)/y~OٹOs}2Oޤ !f\\yMҞEU3ѿY@)aVRtfre)N-&1w˔A%dޣwkOHp-ޭ8Ds06ݚ.n.x\TN=xuӪN[ݲݚ@ȁC4 SSvZݨn2{nt [.[9pfajح-ZMp+*ɽG֢NW(JOoۻ7TnM !)UX˻e)J c=jntJWwoE5hww»ʠ trQǻS<gmhUBY7ޭ,[ @'u;@-f3Zջ5h6noKoջ މ\T 3Q+3amfVL@ȁC4 SJtwrݭ;?rݯTS.-0Օ&kQMK1e}GY@|G*ղ<>:#1B>Z#*NSIKwRMF-lw)X;xݻ/SZhftv5g IQ3kpǶĨ!9,! aA0 Ya!~ߘZKJxe$g] ߵC~Vkbp*mL@.A[FD$Ls4,l@X1I9@ 8y YcaA/ҁG!cQF5!W~1S'ǓG3cxف׋#~@tv\\OWa.L(BYgu~Z [r%79`HjT><^wQ6Kswb|ǯ~\/Z_6|TcmkḻrW4 OʪDc!L\(YudYNĖb~l٭29ْ2ձM5*mL_՘pږTs$VbcRBKaHoLpnwjV\uF}VWD%~2ؑm05gH1֠lnU_oFowx`Z]C0p 2=n?yU?_#*sJcv|#Ա.$H)rN9$ hVŰwF_f "8'aNJ]*n4<t=짡_tPp~[ߊ8)X&%y⭁kqeF"LnHK*vAmbC<~b=Q&)JY` qQz>5vdA 0TNկ;eP}} o?ѦF 3?>L2FAJ #r*% cɸiȚ>*o|t#~eʫ3cRulFLYvUrDTz[X}Pp7U"q{Rͱ|ژf/W:(S jgpl9[Ik@Ŋ*f;!Tij? ^=~Mv\ aeRR쐍qRZs/URFSue=쟤x.T%S`5G>#/9]@y-1\ C-`YX B _;!/)2~8Aë*peKy"nG sH)#VZQ>VPa$¼ 0-FHKaJ^$~|p^g9`m/S`>} 54sغXmkNJ醴:BZ|ݸy WkHɂ E. aD[)9[R_=-:yPa;zeJ{/Ou)pQ&&1F+Di+ꊥI`G?JKU*:GqBQ2 aoF|BHG$Jg"FP*5IFW-urVNS ;ąVmmLmWU&Ԗl|鰍ͥJsTE"mձG3sz_Ж*}#CORtfC0!n[DRLq7r)\uw~kñ*xpLz둋4,( QDp${% JYr\JvrqܑMyݪVXh MTjMlk$߬G "LH$VN\^˺D]yDd]b)w>ţ_/N0D棒kX#$&dYԅB3SiexwisvV>֦۶h1CBatD)ܾ2TtØ۲~Ą'CQ~z%WDށ hpɘ4|b2 BUb++N=dq-*/ ߙӁrveWKN=€RS]҆L[aŚREx]AGuK_0`:˘&(l1il0WyJ$Axiw]*֊3Hj;%'LG 0!le +yĤr|%@ xy[*9Ooc˜9i?#go?*}`:[CT]]DDhBCgaAN[rsy +Bsi]oy>\V]H߄ /'Й{skމq5Y~ K?5"jU_YS6og;2o3벨KU/E$j7.LnvUeDI-1~y[r: , B@ p1>r(7|ջZ*VF;^h~ySZ*t2M7R"ԖR%]j95~ bGqNڬ-.68&Cbb(6E0>uԨ'L,ǫ(92pNÈKMdi)fJp1B&hHXB %{ȴ=yxu\;r:{B[Kys6>0W#0ڙ-C߷C(~Y,.{z-(7+1ɥ5~L!nҁ` MSPٍ! OqkzN:# xoߋWDMʣ <Җ+ʱĩQ#J1{.\pIj\OȒB| 2j%(ǜc^U&p3h"JhŚuw ;U+}vgr\LKq39nl'٩( 6!3m !{8l29( P i-2auBw^ mw;bh#ڋqqm؛h29 F9I&R$MXn6d2JGWas~='֔7lf=oYt[5=5\ȃJyzLM2GB4EȎ|^Ty1a XND&Ba &<2+A>&SԴDQF!L%$WYHy3jF׌`l`@Lư,H" %G>C\+ Dirm(1tHKB!i H1qC?-}"~u+O"Zд4C2zd/ۃ됌ȸ-"2ͪ((x/HnYq&l/bzv[Zeyb+4'zs@.@X*:oipy3}4t6z}rHPF\.'ɝN*qp5XPHavE)|s]e=D L^U*gMҼm].6Y$_Ͱ}offǫ) $qQ?5dqIh.DQaL'e/32&H::%oe4Zԁ+ 3%-0%#p*h%"ۿxۅ\(['#b)u#7jͅȰG-{g5ȕzQԄ G)+ p '$SOčZ9ũ)մTovӘ),d BRj@p7v |f)MR9 ZӕS#))u/O[rҚ:ATO XRp-wY4:P"J mxg-diL=Gdq`ccд#k,Bvh"EZJXS1B.f,#hm0iEvAf ("b nS JɼL΄10Ȭjb?S7䟸݊†B=OH`Ț Qi)0XBFr]FBW w4 w86XB3}0X&oЌTdv3ĞAXnBiH 7E2k@HF Zp`P ?7Rdw\F$vѺ)ꈝ&TJx%?b)I%npϸ,3PNuWpIyga9nI fwE8xg oX;Ǎw4f8JYfeE@Jki.t"CY`w;)j[6ijIsAc2{sn!rrz'qc\ R)~ipO£Int!NS/=Y\az+J;?w݇? uqJ62Fon>L. W5$~n蓻k%eH*nJL3Nj\^ JiO\ӤoSiQҏ~\NeECWw!]4E Q`)xqWba8;gBq\u 564ß_W  J-N [X&ύ'*VYriuq2]V#0")rǡs Y3˔]1onzr6'SԌpr;*l%v$5(x$++;$S lRpt_}q!Mgr7@xD8dkrE®c[XYpuH4|M3wV3#_=-?Wh߱/ZL#ty:_-Ze'דEjK?۲d}2`$;@ 7534(oJH{M]ݔ]ypMkw5| i{:i(Lg}(V*kߧ01r;^}ڐ? UE܍:y7r_Bq#d mk8]iqॸFMţ=yS? Yj{ط_%(#!( {Zys efqpxqqq~Ջ^J́WV>z4{\ ڗ+쏿A-7==ov7nQg^џ|9ouٳ>^ چ<|2A{R{Lncg]%֦uͦ͜Jў=uzQbq6%{ ~heÄ"%`|72FmwQ-Z}Ai/ V,JC 396jp_V3W.J,-[zFm}841EUHU/v4JW]/FjV[UsR6= SJ'p?*z1n]TbɶpPhu{jV!n^UVbJIcfG'[smu},eloJ'?[^gm?FAڨ2W%9ɚs 9.$<#zdiDw^N;1qϾTUle}T#%|F'״QR`CLV̲@@=Ekw;IZG nU}IUݤ"dT @\';\|`#N6g$JbJ̔w&xJS@97.@#l'M1"IEv^>lށm',k\)NQRN^DO`Sf$b}3G |?xY=ysnsiB_9*qcXn?`wzwT J{Q|} o햘^hVŖnLiB[R(Dž~wS fMFM42d`;ڝc.McPJR!(b]oǜ(29Vl:8*IA] 'JK̭idHrVIZkM9 Wdb\h{;XG EիuYYi;MПՉV_=1c,{%?\XvaLOe2顁 CZIVgF8\&bsCpڲ&I~`մ}JR=^ŵv8QЖ9,UhfTj %RCIHN\KuՂٞK_(/*,$6(-lreU \rW3O (\&)K]3HkRR# U^߲@q⣢Hֆlii8l۔SB:^HN$;S"ݜ6}= L1;k 8VOnJ\>_`w6 ?bYHn"/\<(b!p"L+,p(I2e2~,';Ȗl^6[dbdկX,Vd\Rn& J>Ey|ݞWBdlOma1X0<%pFPf`GnV@&qsg! y-Ky*.:G-rQ1v;IRe\ۊ[C):&R482Xìf7!F96mE} [|}FKIݟR0m5J2]oytL[B`GYP: x? pKf iNjW< vBpS-C{#ƽcހE1.$?U$KwqU GYRB: BmPd]x4$ER֮9~Sq#(2_dЁK"70, F`(q7`Ӊ~n?soҚDW)S*NevUvv)F=~8c@q1M ލwq ]F9yV ' Bz<,pYЮۈGpbCk@J){$:e7" xaʐaz5~/q H0`;i,25DNaUd9ta{GYr%GYrvi͕΁L8xFbNiqV*v+| nb kivH(Fe K RY1 IFs&R xT f yeu6Dx{QL0FJbN<'ܢ:EuWc^eRoa":|S06OA$l!TfQ;fdr "Gny" P_O2ikVP*。| d(%Cn(׭ "JH!؁1 kDjA-':GQBT:PT:2$WRplcۍ]{}1?vC߻GqeajK/eDk[K׮z$Ŋ \B97 !TyADUN6Q {MOA6cȼ.6VM,?w翗pZ` sz{PluS++;H( (GL#7H4Jimh)oTK+ C.yIw84;F`(uMEIk4Dq2V=soK<ԫ4mH gL#fCg55<4p%ˍ _$u趞{\5 Z;(3 cdFq͐mM6glMt|H :|K֘Y $Z\ZUZ6#DE\ LmR:G絅nE(`y!kFlf<=/QW:_g﮿^mϠoI>#'z7Řn]2ܨ|{&?'/>M?F7/h۬Նމ{Gq4B+"J1 Xkk'_TpLuyz9p@U^jW8;e kR.Ӌ)D9z7X:cȕ=8/?vt5co ׳zܐ|*#8#@iihA$nb3p.2ɳJpX(kfX\"H>0VJ45hj傤P^1|`9G5إd\aRC?˄j?N6IqN1'n!xE!V( b 1y, dBb&>,? dtmj/tŏW;ZB~6{{24^V_L[unm^]hR-`tjTۙ3;si/.}(kViW/WmfBz=O~^r4kIB^%Mjn<1hmDibBs[E@!n=Vʃ)v׵v+&4W!!\D)E=,;}~Y>ъvGԓϘo=2A9aT%";X \ޝlnB4}n&J)+P]KͼR+/054nTyj7)a([Do,! `t^mQYx' Kk ;s?^[e=+n9npBe9Kn.!z87'u1sRv0lDO(OcD:W VU8" %lK“0_Tը#TP6ޑ]ܴvG~ >罅-0a#2sRѪmZ~PLHj9?CAG$ d6!D%n.v %K4w- [D 6h`locmBJHɱ=Z4|v5lAfY0H8:E~>ڢ%5Kw=e_ͮ-;~c kigYOG=;ۨ@M!Ie+0x M9@}]1YsW%LbVvצ/՘O{}7f5+N,^;XʤIB&SgPm-HzD0BHp\K`[i~7_$i=W՛_|JZ,4㺏g.Xz*ƞZpS>15kaq#qwկTFuV Bwc(<~c0 IM 1U^rڐW.E2=7ڍ3|@Vʃ)v'97VLhvkCB^`| &밐Y hp8wlĔ바^-H+ѽeJD[7뢖 >wx^~xGƄqP RbqFs"&X'; i+$:v{sAjm"s)*rDE0i5w{`T<_!aڍ<1Z5|L=e%"Vl+XqqBa3Vl 6!Fuc,/eW[1=$u1 +z=j2ֆJI[<2?l/&w;UftG;[8%pb;|/E-/ZFhv ˖CETI=_h*]-D"K Q@29g9KFY2rF v6{Ý<5Lip0MNiOi#\sВDMb -B }J5VufLzk)1ia2? SW<$b uwVe'rJf 2ڔ bl=p,E*]Ys9+ l̂}(BmOz;3/@8:MRj{&oDupuw%!Dfx74RruX*V9~FneLcTP|xʨU"zA# SDLKR&Ӗ8KsK;ضJGG̍A{=hOF`i/)c76giME?+S*4t~BT觢,0r?= o #<ᧇaOXS.ᘘ?{O߿{(Rʡ~gL ˛#DWG ~#8=J?ߖ=4WN2*+s*=lғч H~FKzզti]onG?9xg1j_rg^R؋݌̫ nڔw;K饤눊kts&&Fxb2˟&uz0V5haJ8. FiӫrO\v"?.1!d%#g[-ՉFZ*Z01$2Bmt҂CBl-Åp`wJ7`Fw ,޽`x7.4 xM>%562Hi4YI*Y(<-Q6%*0&bЎlLCbeֳ4pQ uT4!*[D͢pS] \iM@ssye&q9u< _~F>_8aCAtr/ـFl#ej+9mU4CJlDhVT7jD!S Z){֞H!Xe⒫|:a4a eD))G9X%vCkx\JU1+^JkXIw"7w0Tш)BbHBun6)S&9|?>0Mf@f=]%٬B)` ҆ñbυRhPB4PBjUa+gb0ewxk>مB^LdBY/OϿ?[)<,:}4x>IiMțEZ˳@e5WXeb@~ҋϓq`tsBT;Fk+6ev GuD.EI鵢tє;YגaɷJP41Tmviȅ͋|ͳ!^/$a:?! e8uz*\A,;!WT0h96հ/r7Pԩ`Sx<D4yTu #s2==MS#BGZj8.4Ʀwu|8ȋI`h\b'lsO"€Sc9crw40slI%5q& '֛Tb5&@oUjO#x !o. gȘj 8Wo&xahr8% =%oW)1fpV%)2DJ;\t2j|Z+&WYCpiu+k- 'hج>J_qxQފȼ&(&ҴM~OfHQ7d,7;kD7؍zR|9iokYP? *5[>3{ח>nD*zM 9Hj+3)>M3HWZ4XfX#RafDMtD+ұԁ`}L&%#*ǹ("k:Cbpp~څ`emZG(7֦ahSgͱa8= ]Ww5eafCP NA)XˑWݿ-JՖ@IC2f! W-+D?<a=OWxa!tJtC0M(kZ)B,QR:A͚6Q5+Kj"C fw@;[%aTV-;t>DZRY\ {2 9ƄHAvE00߶s:HnԮ$_ HJ ҽ\I0mɜ g,Z823#`Xـ)td=ZT+ՂD6 [9zqsQ[Bk^dt7m,oCXݽ ΂Mk3i jm>F]~!fl2њ1t2B(9BwHR./&ǾE]KεUHIb5t5D7FuQ*l1Z.Sd>y~CR@ʡ~53HXkRK.&j2H;퍉6Y%j!^R{SU0NLOa-8KD\`eJ<[mshЕq t95 gN_<۳: G•ZaCH(MºubVwӽ82-O=-٫{Re&' Z9Tm\Pj _]c1o%Y!j$ g>Yf=`Kmޙ, ,{f^ס߅L~7om|4 5s,JU-z"W]%,:Ċ:Ј$k mCC !:w(C|g4mb&Xî!r1v!L1CޱZlAa`VGN ywg=9C IBC<<y7GpxҞeڕ[Gw">xD ;;]WK8 ,5٪:Hֺ:`Ch!#q2Qabv,==F);g`-#0S54}9dB7$Pndm[C4 z{! k qR`^q3h^əY\Gw }c@iP75=I.%qh/I)]Y[ݵ[[q]`Ws(!Ovr#Rnfl}o7JMt&M.&.<~0p'h) ~J?Uޏওke0QRb+$F΃ F9R`` !47>z>`OFS8Ps;=j:og#l[~@*m^9ʤMgow ,Ag]U3LrdR]nFzгrQA"g|$^)WAxG~~wmm|9;Tߪ/XJ Yq}I Z$ oF)p8$e,k8KW׮F`8ނqMɩ+3r߳_͹R='i5Gv_Ε >xUtADE&š:{sn7fAl =<+V } o|!܋[U6h.qP|8ݕQ}cQr,V] 5\'d9Jn%r$hV ޹BuҜ38X7tQmxm ߢ1c{}VYǗ"`UC}}5yUefN:WRh%,䗻XձB,\ݹ舞ٙ4_UM'{ߞ/oi+hڛ*Q`0h5 Q%C!KxZt.2Izydݣv|t>tG飫knr|nGQEĐ`Z%@R c):H#g*RlupՙbD C^;# 咉ɂfª9GAb<IT.:RBHÇ`/(i^ftSQ>1jP[J(2͹C)N&m#'/hT[٫s{ؓv%lú68[k=ә9l CN^ٕ[ +kK=~PGn+3s(o-Tj^p?E.C9Wzl ByK ה%%Y>u~Q=c@{*NBj^3uAX>?)9 l1%Rb6Y*,H$\-)IJ@&A 6'*F/*De,Jt-&$ +dH(Q\ĿXX"ZRZlȊ a/ '_#JCH.Ê:ln6`ُ b;Jހ-2=%vXg\MwO*sh1Wu! #UxvPDمlA%q _цąӫsQ^5N?3hƨj?g)K./8ᲾIZ*4-B)D.OR@*Vq# uڨKQXڐ\RɽZD|^Ɯ{ z~>Yq80+1iOzFz(99vCCJϷWsDz9ekQꆮ^jrj׃ޏe|ŌZv~fn<}Z[nmBixT,G8Z)L,p\|o@\ȹJn r?7!݁].G9Y8[q k;"'oy;u;%ΓTjp - 5 u7 {iϊo׿Ò(DYzة?"z?HYL .S5!æ?)YNYNU0x8**2F崦SQ:B Q+c%H -Ut <&mHGR-=Dɰ l#&@CQ ͉MР#bmB9a Qԗ>0R:9=7-iXʼnμS}޴ 5%,i!ףwcY}YcLں R5'~_ AJ49,`Z:kf嶶2Ҿ ~mM_NfA"1yЫxIto\S~|{v;~/֓0"׍a1B[WZ =6>N+pTΩ})jj)[[ԁN>hS%ִ[n@ hLi|r>h4}4n;"4nJZh] p ٬gR9<ڌh|{u5f6yC1FaRhu. tzzMsLlb2(Cs@UbRAdIha\)zʃRJg+B0oASlU^듓F4Gz:*d\`c4 RӟG{8ZALtJA֞#?ӈ%?Wk_%i:~Z)ZoAu%G|v/J% ߽Ek'ˋugxZa[=j"@q܎)C"Ȕ%wגrS܊%MLWzhb WRNzd4(-&W+U<Υ:4OpXWf?ثY5rXzdՠ1 myS#G-cS)Bh4mOenlpAm―^Rb_0}xt_zvah9O@q# .澠X-?fF.HhA5e_֞i4*i$ fL*j`B. Q3 !JȉJ~YeDz55Ge9"e'wdHƌʝ hD!"`kr<\`̟SoHIde w. GrNYm>+q]hǑg>SY))揹ē$5,p #kЌ#*Wrj J6}e.P#w5a%jU8 Q_PSEZ(!oUЭB-KOh8l0q潦Ň@r[XD!"zCBEYkLc*&*iVE1191Q- Բ-O=*}.{)#뫳ˇxA޸rbV0gXld4|(Svzryr: )?yxc/0Y^ܕ*@Kd:Gُ8eߑ\@ MAONݮjCa~s% ќ?]_^bSyZ˛C`2Υ)!@+ osNemPJ J ۷GQH@jATDc9 -hˇ(d*MKP(DKxJ#e(J` ]B)Q(kPҶP i y|01Fg9h9~ #O ;4]DgL#t ,%ʅÇ4!1j$UwC;.iግjyۉ!^0(hZPC Ns'5e )*?@QF&|2kM (+YT1ohM "Pݯ A$~?=^2BEdk4K(:YS♣)q\ă $])Xu:G"uek[<#a`^oX9v&ݛ<4kpyh#\ G fB BU@rɅpBoom  СT;p2S@'rQA#(= k+tR s@tU992T&FN"[X\ss.ZN=Z`.uA6s|Vuۛ0L`z,qID9sLGz"EN㳚K\Skr.Zz]zˎV50v6,M0gH,j$c*yBchېӠ&[uV$UZtH%nA%'[p7`ϫ.4 L@i S;Č&d/'ajN IFm*)۬6 zt/fҚ'EuEbC^r"Li|k45p0sX0R?E( ЂQuH43na ˋMsGemb~nA܁|UA#j$|4 :/"6"DKXM%$F:9l 9i4irdj}r6zu"z{`F}ͷi 2Ro>X>k%\@z}dal>FES mwOnH_ann# `9N.s~rhьf%oԋzPQmj6~hthuino #g`αd/"2DVz9OAtгLTt$0P":{Pb&> Djf?#>( ,A0Ci>`H0jqvm1>K@` 29\;+dܤH䝣۝ Cʻl)d9_}Qvs}B|-touA_vq?=8`竢o<,~4toѳ\Flx{s?R(y샯fz͒h?RYVkRH lm`dޑľӧl^!qk 7|_jaS>ݘҢh{(+Mf>ʌ5Njדң'-%a-O%9R*DZQl>W:e Z].g^WSZ my8ykUQg+rְ7!:̎SȽfSRGĆIc> Sp9P)Rj'q((d&]GަsnuTQIDw.q*bH%S,|4Jb2aN**qEeJ,A'hƤ * JSkF0' TD;cH5(΄ 5W!d'a$>CHhv!V32# LY2zm*xsQr)Vʐ'@.&AbUg Vk`.#N!tgV|_}dߤ3mT_K9v ޱ +1w򞽅ʎ2ni]ϱ?ξpƢܸ"[gHՄkf"L]4[!&(^\UX(D΁:P7Ž^Q\u=kʇ|j3{u)IB]tLuI6zoaP_uVsrn(8?*{L8?t4t4/ُ&5~w\;/h./pOc'n@9VPGO<6Y•FJ*Vn8gKƂ[6ze6ﴪ6ZBdm5E˸}64_%0 >(&"@,R@&I%6@%A")*FC*p?IۧrZoBV잻oJ*r0U];0c1@y!0*Q p#f|m-&kdXno`2*'8N޶R:bV /OaeFCcn$CPc$ f~N \BRѥ0d`XwRI :.2돧+SGu8f8!UV~ `N+8Qv+ g f\Q}8ٶ3Vv2My|7Ztx}ViFf^^MuwsdT}Ǻ܌4șDYw*|railo׾XW ߜ^-vc& y"ZEߩwR ڭ*!Se /yoiV=9P!!/\D+[; )[E D;Jiyi)bijڭ y"%SͩhNnLŊaGχ*PPќQDZ/(f3񧣘O#b?5F7޲EJ)=44\3cbHccʈjSk (S=J3ZS&Q6Ձ,gY@+'zVO)(ɕź[Tݙ_B3OO"+T,k2/{0yS:+5]nX)a 'gݧVEXl%OaK8 KALg8ZJUȦq S͘3@s^|܌tp+,㓯/0OxM&Néq죶nzkRl2L y!ve&6jr8Y\Ra0oQ?W\hsݏK*k7Xyo rٜYrg,v:fd+ 0"}5[R~/TslY`5T*!s01)ՐҔ2kB[.i3q砟{~~scliDYR83f8d|XZhqgK2լ{y1 #ޢ2강J bzsÎnz+ O g9~.eNȭ{xP|n 흡tfOcjgf :gکa<}4J#V􊈢'< ":0u.B  Y{un}aI0.o{]vuZ$LRuIE3+ىZH) Pǘv {u&} h.-9e햅\Pn$r@ŀJr]9[K KA(4OF3b@C8Uޯ_{Dz[H̝ywz7)ZX$2)r3 %<Ԥ1F攫XDjŴ}=8WwrWo 1O?d{ ]<emn'= ūo,޽%M`9̤&B2Ua  \DsS Z%CORS8 (e.SJO|Cl}[&A9,t!6#4a¡teHӛ@)$$_-b Ti6ī;uhz#ϫ ￾Ep2-xh,rlb/`Ɯta8ph44L8NSzT۷On(?ap)y8p{g=K> s1qF9 5n"Lj̾דq{)Uc#(e86)Y\%HrfMӄ!ƵS@ƺ}fUfJs6jkRjsMhKtCRl%nTmŧtEB8q&ٗB>bV&lOι!u!:=U^qz>䕱;bjT8#M2Yn?Vg-7:}lP&4p(wÐyh~*~B7? _.0ҿN"Ӫ= >b}eq;/ktuQ\fmOpB+&0L)i|&⑎ p7_n 9JC֘W*%$؜[+o5 ޡXR1[߉*Ck;,@Rډ׆?o:n0&AؘPL#PDo/1S&xDmݚqôԄ,Ω0S79FH$rvE_ بn{E_x4>fh<.7i K5l+zwc\U/*ּ7ȼN+kcz1;? ߢbߪ:ix_~lyMVUp4wy~Y@; Q,mf@ %B( Y늈?w<^Ql,ލ dZ=EGڻh!]!VER+/}fH%eN -o(l簋(_C~9|D!?Bg?b͠voTHĵB9jֹ6Qad?ʰawJ@c#au&gStځOз^[ЊczR9ɵR ]kMTh50:FS-Uy. &:V-rb$ng_no*2bl e9"Ŝwc[c3_MnIemI!tOZ!F+2Zp؍j@_ϲ{vgRۛA.ʸ$T2xa.Bܘu7E` @.I2н4&Owqc7G8P[G ļ1R\^lr2iY gѠ0[& J=90k hcVrCp0Ex7nme0GxG\dk,ԟ@2+-z*sEY,4Qj ʑ"V$ZDP!LIy2hD }u=DЁc0тJ)>c&pIlX 2`-Bǖf}WLI/=¯6u'Ai}<džj1X7igl'bEm`*M,oc<ƶs,TXK%`#qE"RYdYr2ep80Lo摩Ѧ_]**bG⿭={\Hr3xѪl- Z{0z ȗ[=znH7T˞Σv u!է[^T~)@h|GEg\@ Y0,!5٩4َ=6*&lĄ}ݱ"eo1c5U PmwUH.%G.zT/ݸ2gi{yTլEʐSgsN8_'mwiV)bĀ*~~وx %nW/M>c31U;D{~B ~2T]DᝄJs4}=xn<@G e ѫD DxdlFj H_% z5JoYcL1; o(1Ԝ{cVi?(:z@)Q_ЮK~k]7J}QCBH"1Zh@ v)S^o*RJ.)M^L/ ɔ #!B1Y~ O@ĂJ G_4AeF'ZŔؿ?ſP32*YZ1]c8#zS3.xNpY`1o@IhA1ڴh+Ӥ}9vuD4ONAd[-Z?Ӓ"ѠEP9hb *rE.btQN3FCшyNZLh\4ܠeѕrapQFP! &5mIJC0YNirZ *t˹7??΢ ȶ NAF+kM9!@H}ULiЫB_ > 95"ω64x< A@hD`cCUAyBڡO58=ak4'ҟ&]{7K1zFG s:7)Bdr{jQ'|߳+96cPDA.R(CrZq&\"ښ%5:JRՂՇd} $yGS=3 5 ւmD1F)mƀR ޺ 0An1{M1p!Y,cFt!9p39p9N( j5iqTD UGNBbhq*FFQ5nk'b3 c4EAѷ !s" `X1RZ{c PٺnH)H3;.e q#xZ! $. m?RVlkT٩2X7A0ݴ@ǂ`+jDB #Rl/cbt9J֬IěgYbƟł: d_~~BFpRV z~}H ,xܨ;V"K̻k!Wț=Ďe?R9]Cn,QI ̠x#V+I+3/ ""߇ɗN<X}^ sɭ"`Yjq 9~WZNؼG5 j["Uo([:1|u;S˱zzc%|#v)7OUѭE g]u'uq/WڑiMw25&ieBS~r6Wz4wwOEo?#~ud,IG~D/%7~-޸~;Pi51SBܳ &0%Vr"+sOjW^>Kʜ AuZF A"׭4 G$VU nGtCyjT "vy 0)'k}r6 _rxx?T$%uKݐ\[5rZ8Ubпn4 | -^,!dRXۧ5 BQ-1J AuZ{$SK.V|Ђiט |28}3F|K#M/fr^mj ٛ}ޤ+Uh.yKOJ‚BN|_JP`Ys`v ˞|e >gtJ{Es~}0V`L#3+fsIr,%N"|Kb iܸBO_] Mџ5Xe;8A5_ xwS* 7D(.Uz?ݍlo S~_K{^%AK7u$DMShPڻK/C+]/PDj^(~ $?Isle!5Oow\0R|= ׶Go&on {|>֎-`ubat zsw9s I7CO3z'1R cO)Ts\\ߜp0te0oe\֙VzjaU Y<953*fd^f~uns|Bf2%.r\{lKsm-/Mڤx(pݜvCwŋ'E|.nUCp'h]zMC YYH!kå탋Fqa2.g:b8tNc%/_D0r*.OxsNUKToKaZE\/Gv]mQJprl&`9chy$n,޽)R4x:PT -YysIc3gLu.!NWZn| 05c6` 593$ !,W$7v 33:OrCkJ<5@G8Ht*MH?}𠒦ǻ]˫&}?RlVL(=?pp?:r.p #*`|M)1$΍>P/Lǽ{l7q#IË́:5X?(Rt惙7qF;wx7?ŴM]& *CsS;!V5k֞\/ =iO28f8zqvܷë+L+%f4WU ͨ7zO+@^6ƁuE;g!u ZVS|EvXs_{.oŅ Aq^$j0 \Y|EhxSXUX])i^):RPPʚ36Q"R%sZs9swl@VMC{@sFKAnJo(ſjFXi|28VXh`;}\PV Y ZA)جeie; W5+9 "`0E`ak ĆZf9"x AiDrٲA xZSE|w2w=8?SH?rk2 k |-yf}HWr~ ~cm7""C4+?6J}0T&B?]cAŗ~_Yk:#/Ff:"`8ҠƤa0QI] ƊtO@bk8o?7xߥwJ  R%Q6Tw."s}D$Rk%-$RF#NHLPlM.H)D(.Od/Kpw PpFI`J`q&kREW&t>K\b *i&,ڵ~if8p#X+Y`Hod1%jK1D6V;cO8kFN IIM79٧͈QSԬsDRNRqc/"U6CJ7>m3&d|WI9&Z[+Lܺɛ6J4UX`$[绰TO&ACe!y7@WPc'DuάIֲܓT;o}!;8NHE'Mu- lHc^;0Ҕn^rAcp[j2diJsABK>sYb[ʛ[җ(' !lƨfg>^"*x&RbEʘYQS OXYˬz_R`mVlꎻtNOILD:h==)=ʙ6)|z^7iEy$c=Aj. *1)/(⫵iΚs̶9 &Yt66D :gI+rA~K`~B]sI)û y`ʔ4GpMl|; D&(.!H2 enKy 8}l`kIu-tw>M(!H[`+L2T3׎bMtM{tn9hd`LHW|%E8/E[_B\h?7D_7=N`iۑEmo^z&囲^.F,J%j[ J^C.yGoHcF̂R|83&AAx)3BAs=J9ŐXX=>gZJvbKѢRypI/}-y&u\$^$GQzp|x2sw?w %гyY4lK/`{R4י\R4].m/2耉ޓB^f͕k T5LHV=3ވ973I|>y#:DSr%1dgQda)RhCafbMWV&>*細y} EAaBJ/̓n[f(z ҇4)\=Q40{Q̯SZchQ'ĈB6Ń82d͐Ls!>f{\- P@4ace o9sNq\3A4O?ˀ $Nxr&dw|l5veWv"IlVÉmq7Y^yX5^gbCAc%'2KF|$NSi<1u{))6sP'p7I6~m e\-|:;FKFɱiX4n{pX:s0+SdHn“܀YK&{|7=gez9NnVs $BkhC`YMbp=r`Yͭ5e5[:mre8wY/~N:b/hm 1ehqBv'KKE%bJa;-ĭc ^+ݪ RtQG&^2ypА]F Lp)W3ȇhmQmQ) =JTECH[._$7Ʒa `l[j q/)s tZiRJm,1ԓr\ɤSMM`J0{+^:>ѭ-Gca&˒\"e|_W Ik4ID' cM(]&s0Kb@~7CנshIX-+o)2Y'?d楘JZz1)E5'3ǞE9*{{킵V%s8jb|_j0JӪlOO1U[lS|$W I=D|z# wnTNp:(+uPnLw:YxIT9f/8䤊C Ix 0DKᜉ B]'0 %q~nKb3t6lљ cUgcqtF0T S!c#B8mR"TJXeҏhle%I 33tI\p6NhPeccIiP L#36qQXedɊޔڪR`&d Z`2)B 4mjL87i!+Y 5&N9ށ7A*<;b&1bƼ*iAuJ_&b앏dUf@v(3aZ0R;jֲ[ϱy29QO)Am {i`dEiUlLq5MQ-Q[*j I-Q;6USR!lP.jv)|rFPm)IpĐ>f>e9pSHn\iQkx22ȏvp{Q )| +^V:U: o?;ZY;{ދ{urϻ=cyKPte÷y:YnnJpT2 NOXbE K[hnr6sĜV)[xS9[rmR)}޶k\6L-%ro=>B>BvZ*~Ҿyӗ߾|;޷*ٍ ?:]3s@ QqQ~zt]>Yi)161G1j\u~f҇EWM&?_%77.?y=$՛Kc?cH:wyPKN_]E*˃X\w.(1ptyhZBmim`T}*ID>% Tp—?T \XW.|@0 f`k$[S€7}BfLOT9qDr2.T$'ne=oc]lK߁,Z`ֽ$d5-`WCp\i1!Xa2/&Pҵvwܿ^5ox{y[ɼ2Uy~z|s7٠`ճl,O6o>]Ovq VBXF!xiKF=Rg6oV% #zgf̊ 0ڀ>sB]vA V ̈ Ihr1)_Gq>iV*4N S3*߬c^t\<=~ti@h<~B |^;viE1$VA\{̒ ՞0B98^bL,x1jXvښvj{5]S&q$d$G'"`< ͘x ZbpNGgEGgEGgEGg]F-:*`{U"Ϝ両u4*G)Q#AA:mT^?jRvi!NпlÔwj=IxqaUVCǚ壪2$!'BRQ)W\}u3 ^xc=ThTJ(vPnka"û*7Q&_@(a h-*J5=\29fjk,9?ԔӒATP/5IiYòF>=LŒg,ճho6,[4K0HU@I.( 2"Akۧ+d+/5-#4q-)8z7όh^ӶUT@rr)5Z 0G0~8FTbyGf3xȮ6N }!#.ix^#S+˒ݺqY[pmJvHPOTӰT_\B[)ܐg/ΖEgylPS\1,vgqu/˘n>]Ee~p~mW߭vfaF1jF+/Wf9e7OƜv\ܭu}Q7Pz--Zpb6{- 'HmG4SVI5~ك 0ͷu6;h4hG}>Hf*ucBR7l)U$>(0z/y͉@ v38(pC2J-mLTf9-FOu)'D٣>l%SVp`MNJheHUHٿ}ݼ} Y9jQZd;컕`Fuew\LjS[PWtq)#ԟ/;тjhk^x0&ji4 Kye}F;[5<{(I9){LۯxIbgO+EYw/=+Z(@u)'|<)[WJlqex.-ǭ2;C2;Nspm4s9t-Wfs)oˡЊg_<Y2kgeκ[x܉I䄍.@0G#DȒ`H"pxuQl@MEiGeᆀ)(n85xÃD2'O$dΉNXK2Z^Zڛ+in#WE.3sĒ>NDu"M;Mt`)H(Z"iIERi-TrZ%rAf"5栗:{B<8|'SYPii23P<M,msPMEH4K_[Tb$)b, dC.`rPABP'KV˜)hmX-:(K4QFXG9N9Ŝ(EXmd>\mD嚰"h1( ^}Q Z%63$ȱ|Y(F} .«~aݏ*WZ "zpJeXضW:8^}~:&ܜROuZ]:vJ R՛دo Zu0Z~mMSH h~0#nC9.:u2)xwtH(FNj%1*L @W~YbPg`pG]HsvNbLJ[Wߗi18NctOwAurITjߘ|2nE1FI -}?10:BT5HGOfI5S9|&bmssD߹;EP/U~ ؤ1\IzI\Ƕ5PmbaZY_^ Yti 0땨c 蒟^XfAzR˥b/&LP V8L;],m2bHn6L}JpiM Wsezb"j({tuf+Rm@]/lp&AF? D+cRzГb.L ` Ґ(i|O{^<+/*SIh$F/F!ѓ<H]&AэE;4(4 **{zjq7g=oݛ ˏXI./ xc0(5q&І%;RgPW9; Q pD_# HMhR 3^sb”# 5$x!pz%XZޒJ--8Ն_ HGp7 `,zƞbnlČ=PrD6Mo>[zR%rch+2a:ǡ}V=9K;%_S.46q.C; E6Rm[{!Jz yP^z ^(cNZ]/Ч|yq9^&E~3!d57PZDT%~)@ λwv\6[\I(+yv;Z\v8ɩZc=#z3V\3sLp!A#=|nڷs^:e")fL1=x\@Wl_>QSZ}7onuf #iID*bDɰy]Cvm[]5"ԒXJ;Wxo)j8cobx\L$&Fx ơIh>(}>=7"GLWrhk#6ZS; ZBHӕHYcOaͮXVZj){-dY\hhsBzلp|wuO`MSG%JHК6"ƣߡppj({ZZT!Jyr^HF(_EOєyztoDb5R-JwLӚP9\[8\kQQ#2ݾ2thP?ec>V$xu>k1SyRWp!5P*;khSUش!$L&MT:8 Dknqu{f:Q7XrlXޗ)KU}Hۢ۶~y_j}_jՍb'zcm>W?$bA"' HƣI5OP7g>VhKXY+]t?+kͩIGuHTԤCQyC*0OW!~8UO&c:073MPV 9ehs0ceNJ/ݬPU_]%=Vj,  '# q,-ҚHUK[<g@5([L30 %fB9Qm k<&Ji|4DPNBYP%ʂZY>3z`O!@YIW$õ,vy1], \-q dZT-hO#[bbʣ6u 2"SG4F놳$P!Dvz]Oժ Uժ.\1՝Av{Z._* &gml9OK7Xz0P3=$D3ZGzso%@NDig\$R;TI@7E9+i\43MsY߫7! HpU eLFYdƬ9Bq4z Wd7{fdFI)NGEY+ed !P_*Wh1O]u%uWUa:bQ`³\2sYe)33se(\5UƦuPxZZsA`yDZ"bE5$z|e+#XʥVznHh8s<"7*KtB 2p1$,RzszBa } |"l8X#DcB<9o? ZEΝ.AP< =\1}|꾯[GaBy@:fnUGz1hoW9y6S*J)z[- pֶ[fϪDb^f\d G MY1}B}{Bͣ`9$7UO!Hf*I 8n"Z.]@P\GD/^fWRtWZEwB\\E2C9|L$&\M7mSqFVyDsnRX!(H, bJLo̗!0xb^)2O_.gBٌaWYpQeucJM4ycz#1ZRaOABHtt} ԯ龸ecMTm<?Al ?^6؀R*%'D\0j`dU _Nckw~=:y!Y˲tFgCf]a8iT]zs_1tcOڃQ"\"Y\N$:h4[{t@)4Amw\s|;ۜC߭~ ͷ?4[5=t@CNZǧe8LFac8|w: |8!껬{?ƝQv"!)B(߳+|.|eƶ;}0^/Gƞ?@{Gy?ѥA;|ӽqww h^כ}gĿIË(LLnt@sLg?L o׃޸_brϏ~iX|؞Rv+u2Zaͬ鿧`tȭ~>O[5aF*d25^Ը?Q fO( ^wd 5M yBw-WZpXcf X?QL2 Qe9ʕT6j}]Z<L@n͝k?lUpYH.?d.!BΘ$3  1LK2Q^gL} ; O8W/ j}_llSS!:5[~uzWϻwwSq3/H7D]Jeo^FګȬBt-KV6/@ʨ1cLG3˥cBlgH "#fހB0 ֆ(`K88oWxQO+9BE}k@Dl+AĶDl+{n!b^S`SΊq|[ !`*ZF:\b4WՆ*zUm|udXOc&DZH];!)oh(MQ-{ ЈBxΝNfMōtOVLbhڙdFP".U -IzGB%!1RJɟnbgO=|`p b&D4d|y[2_%(̡ׯ1jq`B22rZ+y1%AA0^&T-]i.ŒFc=b !0ϔM*:Wy:čw rYO |nLZj/TPpe-{s`{QQe(\0ؕYscn?<1驩vݻשT\"nuQj 0r,!ϔ ,FK,-"A=Z axZ &Y눽04OBuCo ,/<*}m焕T#Dr-Vo?BR ,ǙIQ0Y,BkL` VJ$! 'VKJ]6K:2ul)bi--#V%[5Zk⺱H&T~~n* {7)pZ)|ۗб # =$Jqu{jn- Q?c\uSy- måZʯmC#[y1nm R Tk{lh6o0P|sDeޜwtm{sR."oi ĐUr|!rqnPB&?n0d%o#IHwSG~(1YMm זEIf `&M{XU25Bk}*M`nR=Z_)lxf#,Rǩҩ,wwg6գG{H% dS=zbգH;T3xcϭz21[Q}sW ,!]ЯnEc =R oG7~:\jϭ BEB[8MfňS\1#϶~hݪZڧWmG (.~m.Tid+G !Bvw L3>!u7gpO2I1PnvH1?U W^jE}g3wMN5P]ho|Y/ZM56kW4w]}H7q;FRٴ2EbO^|(+u!#]1E"z]HX 3(>y΃UYm+3"hia xDG K<Ǝd6jF60 fyxhN+  T/c(lPމI+ADR+@a K⾃g~ ovW܉wmZ+?5-@>m:m/3WǍ#9b98}8%%ޒ-)bƶ.[z|.nbQ k7rY(i`7RSB DP % N#F@D@kP#F@I@Kolds u4uF7=MߛwSmq]7ݥ\(߂p]m}~W[Uj#T-#(!n:|PZ։zΫ˱7yFg)|sWnU$>O&#etpNxFW7V}Pg}_f,0̗aފv>m#;KI8 %h%@c.c}Uy1Vir>'CB@ЁOWFs [Q^'4Њ T\_\B+.7i^+1lİV + ]6`a(U꤇L7?|d:6\]2+ޥ -SC>ۚev)'PnnozX mn `7:*J(pJU^"ӊ(OEZu+ @+ ;Z9wPbTn_KX|õUܱba<^p fŀelV7\/ԄRл?!@NGRNﰰs+npzjѭ ?+QC*YAӨE<VB2Oț%2m RղA{) Ehm)7\d,7Lz7ȜSŚ` 78Jo`Qe.Z.Podl/~[*R,yNCHŀdYAI}r nl:bV 5nxm}2[E;.d󫣋t\ BxEZWL}p"wnOiB]U:q@%?s7!(pGXC{w̩e`JyIz*TU"EtA)Ps@Yu VnOi0w0׾5gm噆 s6}@̝'X4FL_pqsC,X M ,9ftA7ȸxG3_A[ܞt*ǥۡ%UOOK>}NldS*j3#GX!R- -8k UtƮC-_x6'=z? p6mp;oQaSHQ>[FyϊhՀL{ AN0rN2jOmAY 8dyQ>&~1M-6Sg4ΏX d+PxMH%Kq {#d_KE̋=oKpp6noiPwPםiC݆ u6}8ԝs(74bL:wp"h"X\kX楢Z!C7HKGO+I?5޳"t\a'p!7Top/nC5]JT""-NC(X+5$xi3` Xl4mWEC܆ q6}8ĝ'lXػ|[ а-0dCȳ8 v먄ڃ;r~;D?*E˔M:Z3s S'S\j`4Hs$5e9wj={{N/hܸqO={{"BTAXA5^z+.'JH{wz,ރEvx(=~1v 0VXSkw_ӞWmR ^bVN3S^m &+\\$4񐍐u%tҀX+ذq6NO\ӫ6N)8EhgAYdMdS; kx@liFU#?\,3yxדZ=J6c[ E5`f_f+lɺ@F>SU}yl\pyKkEbuڞ1^jyc@5zZAY$IT *Wg Hi BZ J dJ#` EY :#a __Vkh/n_{}ы|Z\9vHBhX%aIoObxIw}]U$o!LI1Q̈́Y_Y?ll5XcPlt oeN*AK;bߨˇ׏ڙwXK@uO^f4zxtjjǵ*>"`|~PKCN 6xǿWhɩyZnakg\*}އdȞP Kj@?<0L_򵉹30̗adm5,ߝě,PJP._{2JG;A!2dYKt ql (LzZ N+H>$7$ZA$ZA$ 1kP{[tP:-,H'͙L0Dǂ24 n{BAAbo5=A$"[lǬY?]z=v 8  ylE=JK~~eS@j&mT+[xxY%5zJQ~5Sr=ζ7op?AS*0FŴOD( sZ%ܢ(`մB/*Ԧ-'yеw~mdZrE!Ĺ,\Z|R*zG4(,Jm#fv,/^5^SMW=qB6ǣ8:yBvh/L w/﾿a)>u?k.O"A:ՄWGgKm/1&O~CSJSl~rO+yث L=k —^-;DqwNUwq j|lHBꍿ>жw?,Z]PiM& 22Y%]][3 cF( 2FFk| ކcyo颋KHlE3](9M}ӳt- JWc?LAdl[P\a!&LNάXdD 02 %Hh.Hh(,޺-oKJ OӅL'OUFz3d-1bLcao99&# gq+,%y]RC@VpQ'#*<2 )"P$^$CnMΥ*uhzxqR>BZWoX [ys~y\:D/?ƙ1`I dh#d=|&nҜ-hy1E:dfK?%^8.`RpeeM[Ҭ$r#cd?gPwqHVťIŚy8>1,ʶQ\.JxR䆒 dl 5 M ߒ>gAH*uZm?aM/G'u%Ec*1\dL\pȑL/rH;L%$(KƘL%z So :] B’7L^,] ֲJϘUz,ί_q?_}*rZ~j( Wk<( m~zqOqp3xӇi8=IPk;}z4}a?j߾?<(gOnEv]ܾ:h:>xx|kj ,9#`I_? tڮEǣ@ :m87ss~߼|Л.8h#Tsp'%$a{J  H e9! B& ;V*i .jQE9T˔4 D G weMnH0ew6(=x,cfƱ{V ԞnRf˲'ofu4ǎfL| B-yldH8A{&yHq bJ}(ZѸl|C>n<ؐ/Z&R]jchTamL. ic.m9x%@ڌ|$"-pf'Jj))0IG.”X6w}TM)6F5m Ƕu%ύyBWKh[)M!A(Hy!ʌő 8iOp@(nJ pZ> c>79R,U#"Eo)ͯskI0j9SpSKc @I9P| 5:aȄ%"$2)E4{=4e y1Ѫ<Q/Ҟb'@q'^X%^RhRqTa+}zNXZ35ФD&%U z,XGk$"l~ @0Q&dNӘch=1~秨lIES %QhU$7b(R聋h%lspXs{#=.ڊsR8Kk~ejcNb\VcP\aI1k= AL;̭85-~!XҞV/#j; \ݓf%6ej F״щ&IʠR1*Y[e2:] Mvoƺ0-VmpIB,t Sv\@ޞ1.B}:4b! s`kMMBqTdWN)BNJ']K@"1֔ ih+LnM+J,d(MBG;CH~ː#}_`?]L< VU'Sx~i$zUQ.H  +au>qI%¦couP'*fʻ+᎜߂ [# MF Wᬭ&_i7g dd/*2ɕ#CqH`dޱ6(YfL~ TKMU<םȝgW|Eky'eoI4ۖMM rj 0O-+nCo9Ep w嶚N۔mff˼Y:{ټ{εֻێΞ/yXlOt V9nB]\rs͝m׬KӮO/3Yr+_5F}ʦË;OSK@\浹 ̹^/Ϊ#Nf0Vld?#}\$x>iC"Ѩ+aHG.Lٴm\dqK+Ӄ j:@1R{r! E.M%C6e49™RJ%9WJkC*q_? >J;g8Ə':Le#QKd-`Okv"dlzp}@Fv&Wpg',04l̰]VuNrxX!v#͔0RIVu(FJ$BFjFVb(θPnFJM!*F)I ={k=1C(]Pwϩ BGz^=rNXSsuy `SL/5A|:os;V%+tm.k%8V>m9բw1L -d|`Y b08wzbǓ<쉑H:$cY+Y[Ҭ{*Y3,q4I䒰e4R,gZii1VA0Aɓ|C^=RѠǶ4D~\~i?i}Mݜ,"PJV2 JJփI9ttƌhB %bgYTd>uh"ho[ʔBΡZBI*Ps| .q `KT47w2vY[p n&%י09N1)t9Ij$΀IC@*&U`sfF-,#ɀuu!Ywk"GDP݇e;[Vhf sQ9 GQqp$VϏL>aK{25.~n9۵ >$N4:5ߴ!錙FfgFE߲ܒ &I%>j ro=aSuPpt$DX# (ߚ"5@B<B8.KcߚxFAV4P\lEnI{\))?5F%wq˗U$w _N_n6&^_5ޫ{"NyX\/wfKٜ_T_֋J]T~jUZ.p낮XdT0b˯rȑ`Qv5M0M{ryWnQ֗u=}ٟٳpDj3edN\כ5rĀ~U<6ӏ`q,oc.݇)W;?oz߾7Қ]Xh$ׯ^zu{F3 9b|O6^2oҼ#Dkb$\\H߿Yp cWLD4dN^.A6`Yk! dlHyu7%OxE Qel&i~A/7y#xPsΚ|nq}C_ks|DCHI:E|0?4:BگD}wr$u,ћyWmOoPŢL̿6-07?iw[5o׼^{nGWC- Z"`ɭ >Iom1GeoJ*_^L*<KNg@Ubbꋞ/ {ֻ!9#>Mmyi~/Ç h$0$IC84!TO>%*CS籓ݚs('+^5}0B© TUJ$M!)/tmT}1Чu/f#Mey{|1'WI]t:+#HdͅKb*N>")2 ppv3uJS/%*'䑺ay% ÕE֏[cʶ>iu1BiI %w?PZh'nu#pzӠ.}b<4 M{ئzSl⏓Q%ƬHQ%/o}JR|$Z$w!D.x(]H~\7v>mRi6 87^Io,_aPi핼e^ RH /4esBa?腀a2mAx"o9\\)OD_oȺ۾w9nm?,ZqlW7oWTrh_o =9aEJ=r0ڹ+)om~E{LҗO@ ~4sHV:’Iy|~ S1}|}ub cʚ8_Aev7ɺR~pXډ=^F$P}M nC`WYyUfV(8 2M3FMnxa6Ϝn%_5M Qїo) fVT1z~o'AOB.jDڊ?N9'o'rNj5wǝThU ciIOzr6g9O8^{IpfB Z_ _zV_'gNdUNߘYñGܸW rI `b~^?Ed^ &ЏW㺭"Qq}dX3,H҂zmOPR!k Z[f'"^w5q5u];:<"لgGyvK!K.KǼzfu .3|-FKAcX/wN l^@F'6ǣ1NZ7k/zݏr\q5W@ 6ej#"zI$2Ubr*`I<(p ̢B%$+EHQp"OT 1\#wP:PL4KgE(*QrZh\H™~ӏUZf *,GxOjFA]6^^}ѭFdG3"{f@6+:Yr`#SݎK$ h54u jjt:G~|4}uW`^Tf۝"Hj#([ӽR AX6w̾ϻ7 7>VJ[5db;oY h/ϔG^Mʈ4N{ˉ"z}Co0ªvp.Z&wt?<̊5Q QWSA2n~7ds%~ěSkmҟˍ?.R9' \<fmP12 (V zWM?&Ryq_~F9ouQPW PQI@tSStK|鹜gs 冉Yx#w| /+g6IPmkEϓ\Z.Oz$ .9Cµ2݄Gvvuj)"'z,^i9ȤN1I9vhȧ퍪xUF9!H+ ZgRabA3O!AT@0!g#( x$>z9 ,"+UVzRb0\a&8*lFa\ 2z4ɃS"J&dJU5skb4QaڟM:FT-"w4ѷr*&( FqG"Ѫ.9_뺴(3L\ُېbffϨ{~%8T? LE/|q.%)%yLV""R67 \)Ԏ3ǤW:yTz'U@Y^drB,Ú4T2D{3F>-G>o 9m/Bn'](iG\`Yt~Ds H\VFGe預 V]e<|kMf"3Uˏϣ)5+{FQJ ΏpI?yoX!L9*YgyJA^n17 \!T]GIV _v9"#1A ;Ώ&}h94ҁ|2 &L ƏFx>O*Ԅ-ڎײ}ed[[R1+F(֦R82fZ>Mb> 't.҅8i$O?=eBj垫Qk冓 " z?7-;+Sb;?^mW;F]2Oa{>8}8.:bC?a6}p ζTFGkyYy6grDm}8%AX>e^d!uO OUɟ'HtKgsA{ %gw))ʢDT/Owx^:?X}p4}\g^^^ nk}zz%jԉ_2%0|H eo>hhήήvF1p"$M{,gPl) 1Xٛќ;z^ ?N鮬q_D]o l(?Od>B]ĕmzƟte֦+TƦBwP9- qi`KRڈ5[іIŀ;ѻw4"[ ƺ&H"SIh]D -(~#D@n"5RqN7R%Ҷ $~> q,HwP*I4%7T,KQ9FG ݚ.\>!8nSUPLtU1B|y#j-^[lإv[4) T4QyPE0a[Ůz`mD!E P`kf؅k2{P.70eo2ʋb}`NR*U=Jh Q"Y ;掼A'ʳtԊ5HI˟~ŋ'3aSdq +()~5{@#M(L&h }6k=U%ybYҙr5A!Ytpp@DhYhoS?l3Or?P-oܟ뚦Hb'Fi3V Gs7K{d{2e776;y2+0 [=ya {^TK% wg1|Ԋ5i⯃˕g)e6IF&MK1㲤AL@+2#GF9%=cš2 mN+fp )Y,z{В,8e^?guE~ ^ (܉~'A[Ѣ|`G%RjHW'tvi_'#ڊҒ.x<J kA;yIxxHMxe #粸[|QB]P?G Aa{IBGU|mRqrȉBHrM#Cmr Uk W93F%Mw1IֲP}^ijM"XgaT%܇wUD&]P: SYE|@Myrz ROw}2W#m]r5qJt*<0 ؑKn8x"rSnE3^7ÁWLv|tNRT?P}x1!@.aLb>R@d #l9f˨ >DPYep>ޠ,t'FXSu} ##H1t2F-P c3W#Wδt_T ?u" }e#Q!:%pW:ƽtY9( :2ZPP+׶im?~9k۵*&!@JBN'ʼn8PEPVFp\qH0_OOt0r<3K=_pD)^#/ƌJG-Ǡ":0lԋNK=;0tPz%G[a^k2?~1eWhA5vKZV,nJ; ȧGգ+t:/?]Znxxr#̍+VHNJtDqRF- ~*T"'RC7{۸.v~I1 춃E[ (JܱdzAKN8VLY,Ŝ3NlG9 0!aQeçT.A"VHZbem2 Ջ'Zƞ7MlMwSuǐH.IG\v>.ŢQ>$E;%+F;L E b7@ Dst h[ށ Z=ҀruJ1z\h^@^qK݁3'F;o it65{ʂmzxM69A+T!- g2cDi8J@A(aRR E2F=p\$&;IBR|' vteEbQ)g|; Q`z(A 79pRM&S,LnYYԿ!c1c1:ƪwPް\0nָjri9UHm@!wԶVwv~-9~i}&_yE`KPr^Rn75'ex~VK;ºTi Pk_j 2%rlqL`ނ@#sm$H^8dBF ibOf.qϧESlGLaܻz'j,9o7N̯hI@+7&%M-C\#7:Np@WΓt۠CJG߷ NC}0tJSaM"KKCϜ 2Q hT2S!}ZV |psu6eArCT#O +ևLBf7)HVE=\R%w72إADodxne$HEWOFJA~&R;})FB#)PYka ՘*K;7/k݁ ??b|ʺ}ŝxɜ)K>C.!*%,T"ݸ>"Q(]31K*7H 50||T tqeRvGMHFb dM׹I󺨰-ȶwR%eRTٺ玴T[b9l YOw #d:q*Yut˔lU22ʰK931&2FdH3d"Р܎:MgSsA&F;X`[)r.'.ȑL(ig(8%RI%PdASS#HnxUzpDNllwt}˅Zb>Ŏ)D6V ;e>\d  jay)df_yVr!eH@RIwuOrWL"Y3%M.˗3seRǃ Nx';pFjX\סo-MIN`~׿ǻJ,-~}3!'*әgE˜Ns4p{9 %Z<m=}2f\B~4BUԠ_mHysjHG2!s}&a e. LS,)HfQ1YRcsd+KۛDž g\-=>`'{*A, C$R*HWBٚwco5uD&YY3,zܥy*nor8EW!wj-5dvٜ7g<bV랕R4xQN;eƓ%'Us$o×ۿA9RRh-_s7ߕޱ`Z|Ur]TOѭ)i]OFsa :@AṵP"eo3%x[fqmC")죨+."Х^p>:A.˻V7"|zMdlnInW6y A*&DHP+ٟf0ZvJvď);l%; SDhKsliWE/jɤPrEr(̢P/SdV9-lb.&j}i*$ OSb..>R&8$VPe) uNx*PI-0N4 &V4$B)#; __5ѡzdF78mY!'αOʲ$'$Jı"_~+\8@%A* e1fݖZ>A!. m7>|j힚զFŇ&qCgG'H~y@`k!ߺ\ Vmx3 )S)wmˍks蓥jCgC~R6L2 Cیox큞P 3*f n8Pd83Ļ[Y!4\CJ/%U21v!Ybwu9FrBI$19[=^B?X/U#9#Jư`qgc@ }עŝQT굠PTxkQHߨǓ@\!͉""r8DGXb\ZU49Œ\Q|n189HJS]߫UP*ג?9f/GYr)ͳ?_o~w#ng˼Y0*>??ofi,OSUJlŴeTJg{qqzNtq;͟@Z+KUjR%d.RC!6irs̬ de"$7WY&K{B@+sY #u4"JmuHyd1ĉ [+"/C r@=r"(FAN=bdJnz0ƆZ|hIvrsȓC-x% p{(>]o"<'Th 9C:v캉xCੂȖc̓\T !-e@fa<]ZYvQV J=<ٖws+q9d#pwVZh&,&8$󾕌.Cc]j#yJ'1ɏ{Ujw >}}u{x 3":fZ  ! ߉CPe zFXchpQF 2pSAwLfWvǖE2,h$.S @cb2NiL"!ԉ$hH4`q|t迴S:=fWC&K/Y3 һ;'rn &wtpw#- to!M a a%ˍY9_@ʙ(!]N=# yq} <#F ::_=Qo360́`ݎ0)!lW=!b1ӈn =6҄rvtIDp#wƧHSZ`@"#(=K& a\I@\ZSŎ PE Z#aݔ+J5eEZ-h%=c]ML0Z1e,L78 ɀkPYb̶F2$$5}J^Tv!Hi$i"',C@f2Mu68 =54_^LO`(91(g/ ǩfUaF}ᄒWkul'v\ӽrQf]<+_-謓'q'^U{z=:WpYځ˻Ew췯: }azY@9ɑ+` BEjp\=Uܪ `hf/{Ravѐ%=@"XImC/)f;<3QSh J<Rsd9H9gvH6oW=UoFzj9`xɃfnTj1{ytZgg|12z6nwuJey9Ji KbF믿qֻ(.ʿ>}w)}@/FĘۑo1*.j<|̚Cywsok3wxnv|3ޕqdB $>q8`9 u fL4c EIY!GW}ɩt;u s?^t:< fŗp<J0)ȝf)EL"-֝˓'Nӫd"<]ioٙ ?Fӯvv{yk>?0 #]$ HuT670=3D;_ Çw q ="2QCRTI^h!ə'٧=y ˗BYƽ_kQv2ث~=v!:p8~LO)Hm8?NzV9ӯ}rjÓQZta~feB^ho b6i>t?_`.\[0[])Mӭ]s禥~>M?6,=^e/!O5fM;sC!"j )4fS'1S{N[ -sD\%`8V3GК7EifΣSKANxo(hh蛭\/SNI7+`8/TekMM.T/`Dza?ƤAxʲ&o޼Y )L dJ; EɼF}}H;j=|žn]c[d@Qtrh/{7Qo㰚.ڞŘ`d'9n"N$&cO\mr/?x~tgOBG{/vi?=zqD}Tj B>ϻaw&-At!v~QYT;P/s=M-3ݵStVÞ vSlDHWr^ nK|u,+@a xOK r߂2} A:B,@ZVATj(گT\`QjBrVz:mff4 vI9e<Za&aFwӜyR_μZ0qw] .۱xp}ʽf mu2Gi֏iiׯ6=:L[GsIlNQ'ΗlzI# =aZ.~ dx29 a'YK`Wyeϡ5Tj\ E&P~0¸\ev=i8PZ9PqhKrDU]:D|Emhf3OyN;gmj*Dh.J(^)Yi" q^$^ a[HJ Yֲ@V>op,f|jߔ;gN&c0 z-BC\N^sCg;qcMUyl T;Eу)IZ83&֎7iXuM>赂BAb8t-)yn~d,p;pi|W9?. `0,rc$"/˯ aCp*j?圁L.HمYɣ*N)K \RR/ mu)KU;!J)[ U/KJ?E4=d/Աw\|{>~-uK`$cA Wu]U65y |TX g燩FЭ7t>+ݧҟaQ'yOl`xV^"aBv#k->b#yMeHPbfU08u[5J-e  WjBGǎ;fPTwїn ;24lFa"Tj"=yv&e47Ş_:;w>abU+E ZՂj_X)(K\miQZJ,M%C4#Abe@R66nUN4+-67\?k%P}pm]d9|7p7p;:ⳓdZG U0A)0 . =&8 Ebo% {l o•ؠ ʿ'ZOs% G@ H~^$<@eV[)Ze4c˴T2oCޑAr/{ ~vI %.#$HKw<|@F@&gx%+5dB6 68\,y@%T/+?ydO;y'hw8kեUObU^;%@ p_ @ _3_z)s,fs4). k*f&*4 rƙߢ&o+9F:piWpiW1A.y?V 4}75?I\ 4 U!!MU|I…qE [8A>ioKm6-9:F¤ 03 Ty.H[JwB;;_棝K^MUڡCTK2h\zm`*>nta>V4 oRA a0( v`kJ k=%$#WIc 0p[9myٌZӼlwn+.T}hӧ~1`i}g%&%G")) '#-)$8F *HH4 A4(4P<\ xfZόD3*h/hRB\G1׳HC[" mFUڼ_ǫ\ Ѕ|4sEu K6^/t|EYT{Sy^m6Wq6į]%UZy_1M!dfDntGV%%7] 9ftqpOcIFF4m΋`1i;|AxEA;{>Ov$N=V-vJjYZE~$M7vțQȇ|M3H8z:렾͒ދQxRՑkϬ.rYehΤ"faŢaQQ;Gr,~n fNJxAsWg݆e5gřXOxJ\men좂;a선 ZC)0 d=b`Z*7D!GWI^b\כZYmo& ڜN cTcU0]5 3 Ya&BJ8NE*FdT!p̀*57v[ 4njk a$*X09E\ќs &{XZkCЈF 9꽱{lc=hR-WpG hH,,ҲY)4[1d0 `p07h(h^3I4\9乗Ry f {[u{gqCJ7ݙk|WWXr+Wν/R9rV8, 鳗n^.s-jbWg|߾_6KH)9(sR* w *o*{"1$mh<\Ym@ HG+N LH3*)딖 H+q -$vaPr#⸒Jq@N[T Bg l^"xPLUW]j6Y IsO5пQNDwF o_?Zg6TXn2lŅo7p]ۤpUu4o= :;{.cBvsrS[yH,{'CeT +W G_>/S(n>jLV>ɠ ?)?X/#w|vySn+Q{sY1j'!LTT$AkɎQA#=Ȍ%od^N rG/ygΨ'd\2eԻm Ԗ.$ BUxA aEMhUC е@h/6zfXK` RKPEOO):BpQ*rFK&Fr+-XrdIBNb8BVRjk:N^jS%N;6/jrp)aS!S(I}|4lhwچ@Pѩ>Z/p+Y'M{ETd4ŸR`` 1>$1[LIOx @ØH`w0#D$ѓL%1ߍaoH ZsA'J] )v[᫯vO?ߪn Qv>Xa>(KjW^?ּZnQry^{$k챿XӁZ(<[ <E!:~^; y0:s>/iu)T6j]'A]$%vdaPZYSeaC>i2/>aȩk2&Z#"%.{Im U+t uBF'ۅ1!G{m?ܞzVEJcGFϹ692C:b<|k!]2q"kU Tu ]Qi?`פ8yEJtrE}o6=|k 宧r׿(mŚԱdU:*wƒ>{#ޟFgA;42 T۩MKR?B+NM ~xj`7z7֎}v6|q SzoIqI]%<;KF0L !<X/ toucV֗IRq 4JL&alyhmE5-U͐Ao-I%dUYi݌o]4'(GT&Mٺ42XhuwXWz%G}AQ297k5W^Q+^)ī_RhЫ/58pj⥾xߪ ,>miGn8h_KŦ]جf*6Wx!n3*/F$Nf@VOX}J2ڊlMAh&E%YCd<:ցňF#rayGEo8ZbѣN&<9H?8J^'$o] Glkb;nY@$;Z[*^Y7qs8.,"5eb>UOK9]8ɩaͭ3\nYMU"ls0U ?[s#bn5Y g鋋O뛋r}m|uWֿvf??ZrT,bEKVKVKVKV߮^z7B:Luqŕ~xf%[Az3x\@9H$qzvv1{_^kX= x`lX&Y8M1ݷC x^_Ys a^ ' [FX9Gqf 8/ˬջxC;*%dr.Xwބ*{ Z˷V۶Kb;V z2X} x֐ 0߄sf{z7IH&e\Zϻ4Jg|aUț>>rQDny;"?CX'<c%w p/aß?'~![{MAӶb??x1ǟ.nuǻu哣tw.5Z{(J؛ajjtվCOÜhZzBgI_dfow9]\%*|_;F~4='@<)_xtL>z㡈 FtaG^Q{/5kGgb_F'] ,UpMF;Q}oW)|PadX !fT:m`LѺL)!mNDh2J4f5!=XWۀ%>Xe*%Jm]Q$z3q(eFmGAgW>j%7@'j2@L2rҺk[=>bDj"4'vv9t]aBIBarځvb,ArEk9KO)GiVդlUg+ F%TIAѪNqg;\v2v)uSXZ^c3ʬ^Rf&uj 5d-RͧDk4ց)p,a^ c B` v~0@uCȥSH>#r@˵F_Km1ye$qT/P9Eq0.뾽_/CR˫tvfm[}ѿp{}u5ٯƼc:7yF8{du$=FN}Ȝ| sz>:: :SZ|\nt#6J`}R)eL?V:+j[/̔z^wig7QlO(i֍xrP-X\K{vof+L `Š,I e_4JNg: Lą <{F*0QN</ nv79#ʝt +]@6 3ѪN,Iy>0i<=7]û!,hs5m۳DQǦ@rB0KBt"2e+&R:7aƅ^?Fu1nR phF p7zB2(haK1좖<9?pKkkXwoy WC_5EBu=y/iRryVjss8X0nPjrgKtv1&g ӴR@p`8lS܂Q>pۂ!B{91^}_F( ׺PSyD8,sj1{E~;C;Ֆ`܀>z_[ף..n29c\]{s8**2%Su\&d7򘭫@ZRHډו~ (ɲ(X3D|_=. 88-3]ܖMwP7FTMܙj-uV83CG݈PO%,YE؄L,I@{%€']˱t)ת¹= \pWEc^:k 3󧜥UÚա/kR'ڑg L;z+lFA%U;iԽia=2qoDd&+bQqͅʸz!0#'Rݦ0ǟ>( )K'fŜgoПHot^| Z=HyQh.2,`_9b#}0uO{?j%m=cABж3?J֏ n~A-EhO4y9inD3BQN{:]D\B(& Qq:5=SӣN=0^TBq.Gi\U$|W) CDv<mpnmqvcL&MZC-8a-m|ˆHc3K8lH| ~ :Am͕ߝ|LN6Ů=$Pc&fmhrgÀmB0xkXZ`ar¨T=$$pȡ@:S(}sgָ{4&5ݨInO6 q j~#^-`L{ #O8B;Z<jy)wA_hoCaו0HCI;18s_q[ts9 *`BsAl'mѪ4%6$v=n VYT/E H`"'=BJ\"7˜F$#qݐA, @52Q׌Gi槓Zs1z@p^oe٦xz&3Qϙ"< BB-dV`1')8;Z9L y⭎ɏf6ytWpma6 z&NufZv?[6@2~⾏=>ؔoysAo J}ۍݿrvd[Ht]vzKJwRf]t[GviJk}{t2%#'Ppq'\W]X/j892$]dE1(N9C&yl5(Y^0jz\@~V֠SvĦ5y*p&{rn O<0 k_GAJd'XP+.") YX]wytҦFmSY/| p7nѥtP#C~IGj#-Ol$+dw`t\&9k^Ca'c_Zel1TzFK3:;rݭm9` {B.O{v`?VEu}X-7V"U&rb Γ2P^g| EوdFo&slQ`j; _w &.Yo}`)ƄQZv|˒~t]vv3h*+qTt;o&N[Γ oMv\L\ ݔx/cgkh~\uvK; %Iu 0s㿥_?>~[d۬lrʗGm0m{ A6RMۆm>*f0":ηr; :@3y7P|Oox?wd0 $M+v>|IF#cmվv5ܞJL[f:N%לTsEP.&aچg ߦ Qx~+Q%rysI]4}]S*0H-}{Ł`Pz,?i=KZUdLCo R;Ƙ(4vF7 #oo tWlXTv헩[nDpw@h;fm/R7Mdll*b$ BzS_I1OPHEUY $#"",ǦZ5φwFj+s55"ɡ3ИCwehFa`a'@X8HRs@aٹ1Wƛp4'w6exnmh{ }Мd7QN=|tm1[;{q1A2g8׳װ[,rJ'V6s;ljg|cvHj)kzԚͪml< O 5I ;IXJe.G"}5!GĶt8ٌݷ п h[yDܷ=>~m_W;Yߎ8moڿmF\Z{P*O8ޗQDYp Bؖ<; 8'j}xu+VM69rf%Kuru@X>]khg1Fz$[ZhO} ǘOq#_x7/i %5'0?H%p\[A D;-akxT4Bw]sᛴ7JX|>JXGKRݯMls[WgICrI.fF'J0kÿA"X \W8#*9SorԊ^ H%A)]䱮9O~' ;Erݻ' sX]3#9 j6﷤@?XNfͿڞSݫuX(XH7ݿLd"hO+{T8<ଊ3e8~SMB*0~"i?;Pmh!4 F.RbȐyGe٦x9琞͋Ϭ6cɤxu<֙؟&a9m?:`$a9<3>#Ĕ%\ʰ(Yb\-<$4ɋR~ 0[QgAvʹ®ePz改* N9 3#ubBBF^Q@UuHprR h}ts7XN4wVQA?O&.k;7L{lĎ> ivtc]3U`bU^-RHe*8O;Mp0SMJW*Jf5u\S9fvO0A NjVF.E͵?8#bE!Moɛ7ܩ+wg0oOsVFQǹw?DÎ "6n~|;y 49pN(旲[96mxR\X@ޠG L1%i1SaWD8#n/jA2<6e7甫\~^|Az 1HLqh,xɘ\XsOF0ml>b1Q2aL d8iLײJF9}oWѭ_. bP=@:@sh%@lx rHn>'a>1P-DCJTx8qŜĦV$,VK04`S9|Cm+Ph'-4oRn.E.o' %; N*4[ (@(2TStDU$7LfpO( (X.Mɬ4 FRjj3 80N PAK 9!%6!8 M5Mv:Oww;: Ӏj&h4 3THrJ}فdG}?44< <>(my(fq.N13zla3&פlgf&RU6bi/Yႄ9^Gy,:1z^o/ ^|2|XX 7̝ur^0kϐi RgY]nL4нOߕgV{Ǎ#I/^L>ci @d_d4%-[od"lfHY"22"2Gѳ*} I{|sWRl>{}Pnӽ}0}`)p\K\;u΁!mAPǂQ + B*cz8G(B( s)wXGc^oqL({z؍덙?\7-v=e7*? '32< Dǿ\S~0d!ω-1<5o9C?.OreWySsNڎ])LYHnѸNp~ˌnc.WZ _+^o7Tt˔{ܪ7ˁz,),)/ws> 3)VoaA| [Pp^5ꉄ](}[s4n)*p_̦hyYj3f=KjZXz ݎ?q&)J]!lW8҇g6̺΢fHK‰FtRn Q jxAņ^uo2.6DŚu癝>T9N)1NJzS}?lĤspj Nsݼ?HNl>o=dfT{xUĈjoW;8?`Sa3'V??8N+pXƼ@g]ý 0u  5ј-mq=ᇗ/wDN o緙ħi|=U3ʰ ׷_صW)>N?Y)Ҳ i+ybb zYeBx7? 'чyyoVVqm#,U-r=$ehkB"e2M)JJ"-pSr K"=$n%'&ݭHF}ߪ]¬2I{zCb9߯Wz}yCBi'75Cp^IX ߚLgKSF[W3K1;_|bcJ$n>HiS3z Z٠({[8iw5@T7}Dap!x~R?/J{s8S""'FU[c'!<]ԆV(HX[ɃAEHTlk+Mx1ZVxtNH޼\=}pjn5SܚYjƇ5U>&#Vr5@~~^$x!ORjDwwS2vE_3,wbYbb|{_aZ%K^xqr 0H6jfq[^4"RNuLn/6[xH;ʍUخF`m;Rp8-PAnqnVzw,w;^3rۊqQ뮣fKO_>}ѻµ\xv.:u!^h*R~/ޕT_i˩3rYrj3zL)O2 :%l, `Lz (5)Fs,]`,!@8HO=φԠGK[v7B\h)BaN3s!kK`>HRArEG&ݕqe>~QJlRl֒oIP-tZezI,}!W7 4" "@t,aްh=>4cN­l_'Y8IöO$Iط9ٮl<90DjNE>{֨1'F \Y-/`Ε :Hڶ.i"F}ugMNHnZ 4[mKE.P0YaՂ(2 wU.„ȭ8p2+i0l9"^JHjmd76JwGՒ+DΪ;Ad_O4yc@NTG$ Ujԇi/@VE8c|Vk#$*|в`HN P X)q5Fe%IFC%fTp8#Vq$EB) HD겭(A]jj=|1ϻ UFP2vG}VUeW#uN0vX H`^ 2zA p/AkEqjUEEKnn4"n,2E)kĴ]]ʼnZL)jjŞJ-F w ~଄];Fc!8 Ŕ18#U`L0؝Kv;P:RQ3q0hdӏJf+Ho:2 $ RƁD&&}0GEŬ5Ԃx ᚲSBlFAH$U%+A3LZR!V[V 8w4BV\Qfqi,zi:di1iQzh{tQ(4\@g `xҫzm=VTV/}L-?Uu0Ju^WU}I4SْVzPDȫχ&+~LZ|àw/Q Z,J௥O??Ɠ^xS>EkW;Wf2 ߹AO0DRDbo_.d1-00_qK*&m;"t{-̡. dTSWY(1->i岰Th-9vӂӯ@Ho܈+[aLw3tA*`X亐Q9x iRl6򫇆e c_w (ѓHΡ_vHWs0 n!g n:2wl5y>X98\([Mκ2e컻q3ipsFYU cvse&DrQ4ci_TN}jE%wwsz'8|L3, zx.z4M] Xp^0@+%v:P/*?<3k^?f5Б3,d =ϭg=Eto ;jjU:91 YV4EL!r8*|s ”yÑ>1ϚwKCG5* AQ1,̞awͬSeitRP gc RktܻZ|wݖZy'uA)͐O/ES!Rl+.z4X0/`A0ضտ^SXPB;Au]xi2ʩ>8V_֦(AZ-&w/'B,?)9ɳ4hj⟅M^ret?ӾLyڗ)O2i_go}0iS$ [w2 ` e (( 3|^9I$|-h.w%џg>8*SKhvwe!GpnZ9~ Ao= 1?)2y/Se=c?ņK0e ̉5\( FNzcfQ"+Fi>|o5BDJBDL*c`Qㆡӑ}6 =͊'=lxDNjks1*9 4f"\juY0뚻udv`WRs\990輋G{8I 6"Q$Ĉ<5cŀdtBQ="S@'9AUI\˨`A)@7hEDN !4xR0HSNkD'\[ r0sk7#]7JyM B[qpJڀS"r KAA~XJзe[M ֡Ts{! 8p]no'`}no=N ǷOBER@Y| NMڟq¸'8 S䠠nf H6bo"9tJhG 7c$̷5ssu 4@qK?t *j~nR:]Ľ`ڜC{Z*/]eLwʺWGyJ<mTPUQ(,3/9 a4lldہZy%7OsO!hq5 pɂ e )8++pKy$"{sx~aX/ee,b2WY27KB%R=x : {kGυ WWQ$\̟֮ҩs(ޝk 9!nm%[a)ŵlB1D!{RX3@zpѝHG0J*1eN[ n$x QF/Q/4lB l,vVo\ amX&+FteZGv2l*+uGNTN +eRLNX-SXS-&pIc %qԶX3Bƈ5ΉZ71XDJ<(V$Ha""@T:A]rpYq-rk,P^ ]PoH(~,MLtZ~{ (AIN"qf,,7]솗d~2kk[[b},! ߤhΚ]6k5YϒŰdކմJjIv{?L ,.9>[' RxocשvW]=4`iӂ Y lS`;q XJ K FP߃sad@&BdWGD%"ÙTvjܧojf'OzlnޓÎ.2sT1M/Ri2 uE#AD]6WTŗڥc>lm9$_6@pf"vr߯A(IIc');3@׍~Mř2r-U[De{ޱ<Ȩǰr\ֹPE%dEeDc"i2RY/c]loMb̓8z눻H _Gmx9*0Z!x)|5BKz׃;\u]9yH%XgeZH%[n MqHVh§hԝ9DbKmwT9!%@ YH'c1J3'I5(؛l oYSi&u&7D# jU!TԄB3U%Gh&e?QZ0!s]1_@zͻ@3Oq ~mLvj+#ƨBx"XsZU. 6p×q֮l@~?>zhnMm\עwm P3O=0\y0/O )푢G[?)(ϥFAۥY)Vnw: fsw0//Ow.eqKb mnuɫKZ◚EƤ\_fa;a7MT ե,NV.]ISLU!N8a;&=ѕ o0ⓇM SQݛGSSCkW|}O<5qWZ3,wqf3+.몂 +xϫ ޛ3.* 5a@؊~G؂f:ݎ{& %W\ jX%{suc/D(^zN] .zGɊ{$N]rJ6`(e;D+qa߫TPiMrgk;q.ԬI^YأEn<=FA;K.lm߻-Ez.l|R"Y3F׽l0cN#nombܚjXkп /K8y3s  5џaՏ 63_7uCpP}pb7=d"IçSqRH,HB0W9"r4{Na=?_6<  [kAɟl٧@>Kf<dzҘˋ7ƨrMFǙ[ˊyPB Y2}AH&Y.C {Hf,`\e`)gdf{ޱ ?$+Y1ԝCB/y 5=h?DAh[Eo)ugWC+1%N]5Gpٖ:ߤV#0@ab:h$: {VdznsҔpǴ* LIh)\[5 a,  &) tg!J"N]#DQ 6acpX!9ʅF:'wJo%0fpT T"/bM&ޚ>uz{`|[ e0{[x4Fag|A -y-H9 塅S#NtЍ;z+ E1J "⩅MY p<״ठFZL!NKqh'&'!0K,I .lW/[1n&@~C޼^~|||O~{U觛!|Gzu5(FlA>݁ZOqV7@;3}8z>|`KKC =& )qIm+kkyMܐl'3:ģ l zPdP11ǖMsf4(et҂(,u`=b|1B1+Iژ;D(,羐`04(Ã#m@BʁE (S%Z'ԐBLԋ3<50 hC b.B)5Bzە1Za+d9=HBV r V\[7`kJ 8XKFgyɮV yv>сǿPZ)PƐ##9,9;|~.F(DŽdǹB(l`ꭳBB 0υr!ITys"ݳwm]†h[h*(xTBФ0d$Cw ! MYpsÙ>g 09B-x/*AUnoN1 zCK4rX c5!-qX(a!H‡%Ո&NkeB[%J`4q+x塮Q BbXhM)j)*Bma@'(* Q ʃSg6N21+!*[HH:\ &ĩ ,3bz|,WF( "Op]K)x@!BJb( :rLDnLp$@q&ǤܨnC0O}ȕj_u J4fՅ8ahhc6ZX7RFJ'GTtxcC@']/tؾd h3$vE DDѼg:wnq%tS^[[7&Q`Rk ^EjJqSd]B hc0|uyգx?kF5S_у Ss(C S /46G_z֒/="#?RbWD Ĵ'OpK˵P.WUL!&g,R $VPX*\YX3hP\Y&w23%&\(y<~(?A>\E*8'<e5S pa'<هȾ\eU/cFxxY"]Trzeɵ' 3a.̤ρ\ۘ/WmmduoE>}*>9 =8gU2nېM8Zfըg 9^-AB"]Y`2kpTn0zѬ!γZ4kz8Y|m5k@VwKǽl޶4  3#Ҧ߶hމSz~o|bOB~^#vlv|B~-NC;8FJ_{KxI)C-&n&ESW(%R"B+R) ) pZ`.08$Td\9 )rJ_kPda'h`uMQRNyH]!XJ+p/VyO geEs?h J"=}@QjǓ bPyicB[%^RcErc9=d uK ے4X rUպdt&e֊ >/\ M:3@BNӫ*}9,Wx?o@8EEt|ȑ,Q^# RGJs!˦K.ӊ Kl+5SGEUp_FyMi&J4Fۼ)|U`ok'7/&+^"GBLqJ$'R:mSc0=Pbm p#`rT=Յ5DěYmܲڮuQqlZs UMp?\Mݼ{`, a}5S?_CW2Nu_cݸ: ?i+8@SEᦼLZݬ?2Y1asOjpgR.!_9&TѹmFI [*1&zhUo-yҭ Y4DLS"ZhPՔ*6K{M-1ܖlʙ Ҧ&51ti&[ڶ3ueIᠨbN\zЅd ţ(&ެK]zfunVŋbVJ&ek AhtVN&Ӕ I:ILpVL㸠\CT!8 % C  &:E Jd;(yFmq%Uˌ ,T/B,q|¢EŪ"z6GWso@ UCk 0/%v͝%, dK7Xg7;aե,czVsM:'mbIx'.eqiNbWIQyX_6lCJjI*yu)X=3WJ%fжh[6O^]khFc1_y=[RV Y1Z4%1cT {<rvo|t@s|@D:Oބqc1^=XB3FS!ٻ6r$%m9`da/Y&ů<%%,SX ۭf"Y Rpȃfm yfoNfd7ݞA@C QaݖL8U~p1`7P`KV*\2,u*2Y w%h+KX% 1MnPe ÂV=wv RW}ߋrky-wʗEW7ZB g"@>SٯM>ãO˘^^y@(4B1c0VJ2+H(){or%x^n<{X?,%?|(m㥸a+W>B~g_yn/G@Q+z;2(lX @/)<4(l.6pq+৥,s|ogSś߽t7vrt=[,/63\}[0`_vUQƧJI6(!<AV lӃ,7S'A$ [I~zXF*@1~K5 +\O?L* M"VPkfɊuv5s?CgРp!n׏F7q7-iGo[b1qˁtvG tU,nv;[6c0cKWM+G|(ݐvh.XtU!;x26 ^#H"p=NԕlyQXQ'P ]ZVa9]PQتm3n55n +TANͿYHRh l 56Eش Izm0^/]mܫa--4ix(4.I} [wQkgy$U'o79򳽭oGR_]͂rn6%%G?oxVveco7OGBVHMı2vrI0:k#HR yG!+ iCJ/?Zyt}wAQsuXF>Qs5[oѨ_z]P -ڙke Hd ktۗS QpfYi.Hܳ*]g-eG\ Oo% @lj5L]fV PDPYs eZYa7TgqpaϨl[T3F-3r^ga+;![6ErDzZ2Zݸœ!]݉/v1w]XNS:AuNJN}/돽4!J9N j6'߭2D7z8h͘ ""DU_ţ?!_|\.6, h*ך>=y'5%ƣkXuUx>xGBߢ w+?<^=2Z_uh["D?H(ƣf\ e/!C^Q' (LV&M;} ޜJ|/bhqcю1 jY"Ib2RWǬGl!T*V~C.r_űMUb,DTKR̬<> >ŴfM xHӁ%2L4$Lϥ&@=&cl :k%HZV2+sK_*edw*'Lb@odo6#pu (fݠ]o;ޠ&P̟ڮ׺FX̤0GH kS$,P>p 8ٓObpU5>~QNQS*h+cI#T${ui\_d('uY|t-߱XɹΛ$.oȼ\^MxQ)9 lYQe":r=z +daTqrQ22'2芕(MowzFd[cp)l(s$ύ9꓏NFƳ qUȠBǁ61JTGj"~UyZ8 >3s>܊|6X`Tzts6a׸uV~60 $qq|g~}NOIE&*6ԞmJR^H!u+N:P?@LM_FГ,qo-.fkMzMzMzM5md>,1(+#E\L2A1Չ*N4A4q* JD3,~zuX%e!‹ ޓޥ$\wj ry`3aXA/Yl~*Y3By&;CG5뫫CciVY?sKn$!1LΎu2NÓ*$Y눧ƮAJ흽"rFDB {[U$X p'Z_u QG/.^@P9c] ?fOfOhх>5OmRd9,)F7˨6Kn}'oj_ڄ0%23, IĈ83I~[]ۙ-FM0e\H|.zQ:wV@!+Z/v^gvGcqHEh t 6RZCJĒ(:`0IWX㘷y;{2bHPBP{Ɗ[&l+G0޺̹0S\8FǝNfClscquaW℈>Kk aum~ȉRAVpgo[J?_ybY*B [He9?"n[\e+ kC//*TV(Wq"Rz-[xŽhQؕ?Ac>$OX©VRz V*}*6K`Ylg` 5* HȂؤUAV8=irVզJ/(lPݴσ,ru2ـ#$^oʌ^ #8x`z*xP0,=`!m;PP i̹8G]^˟܉ 鋞C)j2caЩˠX EC(m?B9̗<'kzI # N]eBLWҕyI$ibBiRlycX].fV>ڝL`zQ`*? &b&[aS-Rem-VEp +R՟"j-wX $j%ܫ6A_dhGvg{ﯲY:dy3~ ׳@xjA5x}=[v;Q6wgS cL^ T0S,)+ %:^?aq%hmSZއ{/U{*rk)xϋ ;`x( P<buwzX*x]g xfpx}t7{Awڲ.Wx,onoϣx{gFG;{1wˮ3{ݯnnչW꫑Ղ9dᔼߣDnf)P~zXc7'gӣƯ? :8Ol\+JqiO>[tᵟ}Zt2mB6aDiDʎ0z0zZ<.1J'.! ű*gQ;)K)p#B% ;\viCʎ4m.SĘyBy3uOg5٣u8|0{W BO72k经@pO8*$89>?i|RÉU !0K\A4ƂiR؁&SK- v>6lxmְRއmq'bӏ2O*S?LJq\ k%Xb1J%)̵R#X Xa SqLeӆ˦;!9G2IvOBn^^璄O8"Y-`"T@JNqYC|Z#7FJW; vҐ#Ȧnc6!q>!6.J-dYy0-GN2L4$5.izb:h};RjƎS vh"gCI &+p H"m|{g&ZԩY0KFs0QAF8XP g9"ZlI)c$ ),F`%0c|PqmP2.5Q We[;v(x*ԺH/gX_ ClPE-b;u7_,;{'!+=_IZΠmKULMM{aˎv$h+[$VZ \Ȕ.hRj0T4ϱP,\!T! RS AukrA`*ou>q[c_Mz}(f[f/m919]֕c 1g%: 9&v!zK떪yfW%#mWwc(;jG{&n0WsM,5Co@!^.A~{|s2jQiNHMj!AMm4"UahՋoA`Gnq|`eH1ҘqPj\szظ\l]WOͿ+!vL SBC\`6L}nۚyZ"ut(=v,)oЃ[\ph'<@X9ݭs^ ƻ[/l55^/tK ŝ|aΝY7*kU:"֓^:s}t>Y=6jw\ط(Z 獰ӝsx،PaM%m;3|=0' ;[1Y‚DBD໩]DQwM B,*L&pEދ9^3O^TZOVJ4i1ӛ#ˈivWBM>Ϭ)f,ۿ}cqhriXdyɓTRL `̹m//VfAz3u6M^\]IEJZd}C q馢Ri&P 4v󛭙ffPd93Sbn>4H7wnckˋ镯}MϧOHUIVzk0Ra;9Tn ZŲR+wa%Th;Q @DCm`jWN=YLUp%[%L6$pU@*qe̻ni<`#]#`M0XM×a@UYx\]1[c+N>hciu{Mo֖Xo6ʊ[ B&R@AQ<[2o{YR^yգa(vGYc NDj3lK~dqm_j|ҹ/݇@t⸹aǯ} xL>O !>7^uH)LJ2 Ʈ+\c/AV<>1N '1'M8 Ɲ@\8E#|n^^aa0 n#) AX<ڪ븄?^qvL =G>@{tkDFcF}w | J$$7񺄝d8ujB7m=4OMXsvyg=/ӄ#쁒:-ϗve.,Z̿N=ǯY"$E$G 'Je,w\)SE>l ~'j>!˩W_&kfV()&-zAc6t$@Eq-ɧ&Cszzy<8L{xgu>-x0iDK z1>M#("d>*xw^oQU<$mw<ҹ }Ϯ+V;QI΢˧;DQ0* ctKq#sW)Dtpgݳ[o)֗ȅs((_==>56@O5٬TH4U* $RD%.i}|X6oܖi'|}u}Hw03$rk?e%Tiْ! $-q4alaqU!$QDmXYϚB.ziBqW(Ev Ѳwot/`,FG3Yo~d:=qG g"BSU֖qh8$XQ1䀾w{9㖏+'h9 ='Q9f.fܹiF6M$s eˀc6tոچ+VW*^ dU5 W`ȥ&w B 40*xXB2WAnF!WC1Qa ,j%xJ #*a*EvVzS/v&_BQn''y4eJ2*թ 婦45dE!pAYn$|Ҿ誃0i'ӽ=qku@Yx8ꜰq8N'. b2Cɒ@e87q%H.C9@S=ng&tiL 01@31t2zٞo7^yoҗ-ӓi9UkZ ̇>%1rb^~_?~å}eFWOS/ݣqऺ|Յ LgŊwOχ/ @brz@$t`&bU?WKLŭ8d f}X޼V, {0RK!ʷ/]{ B^N Gm-E'rSqvZ09 <4a6>knNպL[5ߪ 8=u<`5dt_݈I욀V]ko+F>}ρW"|(NZIڢhˋFIg+kV* rp!9$j /68 KxҐ˵9 qp7Z,p8a^⛁cjr7u?|[pU_]Rg?vo{ wVғy\'>cI3>}QOQhuYq?{WꦣgCUT''ҭOy>-1ܻb?0Vg{+?泥[9oe.ˉZ(&Ops:?^u@z /h"wcRo֭;2m30jm1N{GTR-{:\,*PUGZQ94>2X uC%xFYYzpigCXDIXDj)}5Ղbc"sI VF fA(SLX-45oΌ,E僋XNaLϜeștzfU]1E=eֻrƍ ffFR{ZEyw@̲g 8yy>wypY+ۻ⢬O~>YeIYeKxi華2oXKcbr\Ou {U: PܽpɴkLZ) 1 m~k\VAяq8-av}I_Xt ݙw >Oi&JORZg?1ظRҊZVGtn^ܱ:5I?$i|_dN><`.x.lj hĸQG#Fid$>Eb.x%7BqLq91K%qLGD}CX]1pR vgYrDџ. Hc,1 4սAʁ۴sӳPmwwD{g.M6wq*1pcX|>uqm/øJŒN6rw. f@dd7ƹXϓX Y L$`=#8mT е<]U73:U鿧/_7o_:̆A?u;=~C|ɏL~~/?;>QQ{t @ce3egŏk:م|-[o(|Zf5679q}(O+>ro.ecY'|^{P"9JKtqӲ0auےz2>'Ct'_ZwqU[psOF(4{K 0K5?Nyl5&ۆmU:)EvuF!n/ool7/:6p_o޹U۱7/}Տ|@3Ύcmw鎅6}=vhZxy5\ycpӻ|:]GC /~w)HoFc;n=+dU?Bi'NGEq`/6 Ե_(Db\ǃ?{t;Swo1!sYd]犌L=KlPѴF/7b\z,8/dr g=⃻<9SXia.IJtt(ڑ9̬}P-;IJ%wFׯ2a>EQ8@LZ4o܇nYe!E._wzW\e"evQr 0D$$ޠ*rx:`n~Yow~K>HLX:${ Rjj R沙 5elQ8!X6Eya]vL1շ0YrIsk`Ŕ=G: !1p0WOF߽{W]|V(Z i, bloMU]}?+qkAߎ;W IU.u,@[H+E|갓ƃM!ZqL& @ؐN+66"w0rtfi-XLDH#i{R0s6b ."Tb1i i  ^ 9u.ȸF;}R&j%c ECĞ*Ĉkpq)g<{@$nGix%85q^܌>u]?PdZ `#>aD t4ErR_6Xc@]1`+'i@Jf"(j^Lwy)C`:,3ETF–E,Ik xDޝYMHc`rHXf2"eMhl@C RsJP H8XrI~CZBBUM2'[@ f6EBDw;c}[2GW*8*y"S %8pXIK9eް,Q:Ij 6`c-a8`0e}Gܧ⅑46H.snILCQLѝymxG(qjcp\P 0e$[ԥVPlMQl7WTG"!H4Ž* f" sԕMqG|CH]kV960a]Ώ`MAR'|@)fQ u@Q!WV`OȥT{- b-:EjvcNT)N@j##ɪVaGņQ@6\AR`K`a R(D *5TwS0&9JN"Bv1]-Q&+Ngr5ϔ䝱FH#4' |mEhgE/C")yT3%vΟ+ TiahOH?O"`XJR-ЅOKrymX<+PEJ!ebBP_&uy4HǛ2q,X3@j9otch)@9x >Y̳ȩt#RGJj sV[alXZ :9E0Dq2يQkI \\cs)pkcz71ŒB*`lb0ީ!S)a$ A:4P-S%WBJk ze$9 NwXX'1(@%GTUӈ*s.T'Y$@o2Y1BQ~ 7l~D1γrC$ CB]Dgx+fUTM)߈sOL!wKg鳇[u'sŋu ]otaFp+ZNr-DdU2( dvpY\py/W0^-q1E}[ fY,_Dܡ>?>0QW9mS+\+𛝹6BCUS@! x< *uA9[[Ud}"),X"oTR!+h癩}z;ɗB?xӊG 2N5uFB;RlO595dV&KMژTV*dj% !wAՐ~r-úVzш7]чk8d -hƏ,b7_ZE=j5M jB!LĞ›a֠sҕĄ5g>'4\^t$'$#9h^.MlL5s_EiܘUH/j%R?\hIJ]2E^?Jqk9Tm=^E6-n:cD-Lz8% ašDhf8OAiJBNPēObC,2_koZ oUYJ/K=H…"ǚ=J?|rzWyǼw`,xIRd}BSLeJ1E,L -^? QS-TF]t&4׻8JycW4Y`!sFiRL՘ y9p3WXѱY[ׁCt&;dyATѥ>[6i͌c h G>X^9p{I0,4)2P*+gRƝΨ *S0ۜ;Q\y"-b֨h, -c˴fOpUcEk6,iWZWoI4絮j!;6stj!O06hqQp7gtϹh~wYIY œ4yD;ݘ@XӔ$ [(d9SBJ ~qŁ0o4lѧ7,7TFsKF'1Rm? UAO0dtk)#@dFG6C Q R8Bji`>ݳu6OsV8_*HJZcod1nktO*&jcA^ؾ_RА53 5@=waR&H ]4O )cM+_( L! yݼJGG't{^w5I )&wdƢd(a(v"[8SX>u!uf} iL R H{F #Zg6 늍A B\-IPHDI2Z,)c8bS(Q¦P?b kiϿ^y R'cz_v΢?e&Ap'Nrd!fƞd[|ib4%vX YB80pF!pgocMT=X+i՟c IBE|dqmWl]P/')!1xk@~81 aK( ,/e/F6tEH= /i#ԐϏyQ7vjG=ɍC%|qD;ExcW& 2%mDcIԎK2pII <"|u IQDU7h9M$Pm\҂-9#=Ʃ|'?ˤ"m% vdW4ԲDhCJqN⥩ot)'͔oMj,ē!HOc2iqמd)LDHt"Z1Ǭ t<7 gR bJڎFov~3 "P.Fq2>Gh;ߝ}yG,$g[ ٤hޥW^uӫNU+0AV+r̯mNPcI,^B}+뷭Gf}__{9[ P1Wwߟ4&iUa!,' 7i?}sz6]?޼{ ~KNj#Bϼ˧[vsy7- F\Ҋzxzϔokf#ÅCKqRQ+C;&D7BHr+RHG7&nU6 qH >yB@e)i&90iMdhu{։ֽB<\џJzNTG h@q3&% DD3vg,2/LN*R$2pc[Ӆ<"VR5,34 BW8H*A y)LJ#%N| .)\%jF%jn$n qܬ0VhN2nT @Z+8Ti8BQf !z*!S1erse)CF6kDOr}G4M8y EVܣ&MA/! PCBEJ+w)bxmGUNkCC [|oM-5A=E#7H|<*xLHtJ`JNz/)eeGuakynqwc/tLI1i^xT#94ьJJknh P ȴіExPGbyPZe;o:|?tǞjcfZc0k:@47pMɹaÅ֞&_ w(Kp[ p5 ."U2kTb]1''UZI@Xa\3UHTK+>P|7k4KWP~ũц(rij& #Q0A1r J T5AX"|)z%xN;ʢ)ꆑ8լCOgs1++U#ZJhq\w#WqUPjCYE6VU.agpQRI:ȇZ\yLVO|W \rsGYQ5@j D do½ r`EL ]g}Ze#}]Yѕ%Qt(ֿ_J-6Ȃk-}v;]Լr(DJ~B К!wnZ>_0N^0sIdQ̭HDAL 3f8vI;h͎e7nٱar@&^(%pdu&ae[5>h *pgvW-[7G(j {n1E_\;{)l=y{R\ Hݝ]\<)񳏗IЂl-}ܟ|/xG{QD`+|k9%w[ ˚WBzt*,ekϔH'i\[| fr!R[8`f{"PZ [(]h#dPv %ˠ^asp=X+8{p=+8bfZBb)|w=V+$R@ h\-=:T7VKHB5d~'5 rCޒax'(|A4m - փCAP!a!;)l=<}PTeb.[gn\IڽmPro}G(+$u/{` v "t( 9:k}6|(Z\V2BFq6cM%y sБjf,hMD"+e\{tɒȚL9pX7YuQQ!A2%"Z VGQkP-4w4:*Sv_Vܰwp& P2?-E&ݢ]s݁_^(n=~.8;-tB/ħpQQF~XKA7.xQt \qL):煫ABqFdlDuQbpY>pxt_5֩CɧKk*rZҜ ќ`^.)43jXL)hTat} 2X{^zޯfzZJpҵSqȐ?{NYX똟W_~ k5ȗ}|y s߅{>ѽjjA->l}HѺATҸxr]O-]/8g.6h k-5?RߦoB<'U(O5q<9QG@PF4UI#4 ".rf" MCN5h<ؓh~_}zԌ4 VInoݰ|븫j-$lL_N~@$鱁>fm}OS&x2L^ky8+ezXp USwOPR/;zI(8%DEv$ tV2E-xj}. wʠ-'i01v~6jhlS.n%pgu2gg5{w7~// {<Ę &vx)mU&"skg!$})B)wKyja5ہe/+2\vj7MJ[w6}zwq^ջNmdoJXuj5 M3 |u"Kd Ċ';Az>Pױ H{m(DH)D2)yִ y-ZGrXl)DYB &08&XLvbȤL H}y-&u64BZb @3N( QԁBO7wjb^l7b` x(@ƀc@ (3F*LR 4NC-`Jupڔ!hfQ ϗmbX7WWuT &Y 4 $ ]."u#7{5Q7uBYhdw#pC !~_3Uۏ?~Nɽ[N' V9n]qP-rPS ="w KܫX9}Zs0>DRBQ!"AUUڦ((NB%tIm[QA q~]RIRn&jm qNB%tѸ֙O;6v%^H+4RBQWS0&tJD) Eщ^"װiEł#Wi^e"Q"Zۗ %/TBJW"[/{Uf_aJJ{%l AwߍxBL[uE6ݺgcW ֈo'Any.) ~^dUFW\,|ꄙLItcJto;^r&J8-Dٗ<)9{sbU_"2V1BsR9v+ODK(řAB~{*<{Y~8#UA{J "-S!]{Ja'Z!&Wȼ+1|-ȿbЋ1Nd-oOl8;x欺4ok8*%,Œ?^}Sp@Զ "v$/I߃P†J[bђ)s kUc|͝ rӲ:|_Aj.y MBd^.QnԣǗէȯn2M46LhѬ#Gˉ7!|<[9^:):(}fІP,<)t, 7p,'4(>h!Z;beaBřplU@kD ?XG/O?GM0켥xK% WMRDF{BPfdΟ&R)! ,M ^% WQI/4UE  SQP|ND-q_=;xq2-ݧȋ} -2sd1VwWΣ{ˈ8ب+"#L_jRN0O%Uҷ_LPL*nZMWgڄ9R(iJq:27RZ2e`pe 1Ki*w8X#]7J. x!6$E)y23 `L`]V-č_7$ŮWq ]΢WaPy|9B/4>=UӇUq{}UވMzDd Wgb>ξ-t%\BJSL(}67@en`[5YnX@ m8Y!BK 1RyOBY nh#R4"EW^ ZgR-3HDv9CMAх' նK/*Ɉ"0XUsq>3@+GdQڇL3Tbx%( SŅN-pJ,x[1 ƈr$eS ʮw`wӃsq [m'܋GK{ F2,82W٦TbahMB˅ӧ_J0.̅X;Mw ". 9%M,!C$#pgpʐn' ޞ%Dz/4q' ҤVXJ 3_"AXrHTaf1^)'&yN \EsZ;5+aE/WCoA һF˒ @QWܮsҝIB} Q-=@yOR;[l9uFMJ/9@dX#&rV!!DBB g2ARLy0%:iu`6̑RHeB2l H#atlzBYwC̑^n[-].`o O"Dt{32bb#ۼ.Lr>s2N1r0_$ScdB͋Q0l$BtlԐR8 #q.gdQȐN\L| Gڇ "{w{;y-JK̪fUb/p7߽Zz )kX`3`)}29!`=Qqt?>Ľ׹J;-G?!dΜ^|67k3 9qk7Qj(lJ !j1Lt\cAIS|? " 9Jt\$1]9h=ZJ OUM=3%g87Qs'vǡ }DUX2K#WbB1D!m:m$ǜ_VC`7$gO73!tJ a9׋a(͠I qӥ V( {kFϣ^&MDr>S= ySloxS1Fi¨NJ!@h&,!R+lB(r^6W*WEf<(e&zmZl|{ps0`溜/I:QzdOROOl#ԍM@}tk-fmů{=,1(4j> 1?R?4LzPhzZ%>-AK52þ} &wHP'JOXQް:t쵼!EQ3#ۺjH>r"wYe01 &Ґ&!P ea;ElT;#x t\ lzex)hDr]9Aw^q ⓞ:X=eљD9w"c\Bf{9,&r WDXٕ6Q?oIIs9Rޫ!߭@t+'_'0#ّ-z?0ɔ'xr3rB9o6 B-#mÏN)#Mۿ-c0xq;=9FƎ*% P Mh%ø! 6k6*_pC'%!a)1Bs3gh wX61@N"q YQ>~_Xq18^X6Q(q.<Ogv/.$}xgGfHDp܃^U9W5tjq' W+8D\UH͝#ddSVקuYlAg"tJ\ti7wG l\m:d){WMivj.$+ }ś:(c!'Iԓ~JI93Uogh58UqB HqQssbZARB,Xx"$ WQ_]%߾$?vw|H) \֜JJnYjGrH2Q6c@q_p/ @!eZ$HZ$H>V?'00xa)ABE4nk )Aj9C"e"@(* "uhbcFh2 B&_Uz]r4 #&Up4kXuۅĽ Ʃ8igN8pM)(](<06!K830?WCNWXGT<TF~7wŝA"'ݜ Zưn` 1mc#~WG4B!%]1F! \ /miRU=h55+gg>ɶB\>=ތ'OqhO},E!Mc)TRY,ckT2_]Uxvx&Ҩ<6Zܺ8`pN5 ᨴ\59"ݠ0 J.<>IMSUg􏗐! A` / 3PyٻNH)$'aNj(szX?o%1EvK:8_<|] 8XX&/rJG)(2@i l("n{w! t𣧭SLxm8N[)*+(q6 j!Br$RʘFBb2J"FAŌV%[(̌PDqnglfn"xU(mLC%YB,"*ML3`R (03&)R[X<B*2MR"mCp5nKwli(Ll߭~Fu=;o ]^xUѳYs{E}6/EpkZH-8EFC2F;j,Kw:k9uڬ 0`p!""? K-t AKp\UF:϶-M|07[TӒ[p&Ij#%pSTʍNq,5L2ixbG0xnĶzw{40c 1D$"$$吨LmTe #qN nlSnjAj%+swa: z[ ec,,r/Hl2(7U@~!cc] oȒ+̘v߇^ pJNߍWs R>6Z.XIbU WT d/p1h6%=Y B5gk:C.6w>B^!cOo"̵`gUE7Mݎa+ 0U9$_-A`:l(U-D$ZaE;έ1]Vւ6?}[ܑO'{6Э1Wv!89s3u^l+N298te0M7ml h<LXyOFupdF V9K2._ +g3Sa4K+ 6PYr͊0$X$v6 䩪<ϊ\l11Z pg$JvI]*kLHLIcf^IgŌtzU|GX!y3JSsqe>eu\YuEm~KpE=++|ǝ!_8RZWh >*Ȟ3Lzui_:P/j={"5a^m%9!Ut=/'&A|Q`/VPE6PE.qX9.YҊ,mBshv#'uJ!ϐ+JKNaA5 TM2up:X3\^,療B%Z]Nf&)76LRqф@JGDѵLiulyؽ &{̼3/ <,yC`~fkDYӴ4f3D쀗Lh.6k>مGٲcR.'ڙ Q"0!Q9"bp I} l&Ȭ۟n#\]|-A:SB C,mE$|IS(LZAXEM,Q$V f!BȈ2JHP0lfơOY JS8bmu*  .cZ1FT'XsEXjDa02zq"پ,AVgGjBNvӦ1eiIxfSL5ERv]Vkp8,8EF7`ٴсG-"I>_xv koUiԙvy;cefvzY%@Azg+A;,ևyέa"5% l럫iJ?V[I V uA' ;^ dy o?^xVNʏ#)Ȟb$!0Ll2ZV "HPbk#$dxCb/m:h'n5Qe標3=yraX<&;IPKጋ\2 4-p|R@HեaN('Aeu*P3ʩ5#*9R e6I:(\~-C"&̣dfLWAZ #=g'tkr0 jA;S9h&̾b`E0:݁3oᕯPIzE0!z4C4,rj)!q@7A00kHZlF7czN!_ ˕84`@t2q~ޭ"wtyIx!s8A{7lۿ LW!~,m;s2 qϙ't*{VL$ܚ)Z8@FZ(4W _s _z'w~=JL}ldCkŪTGϘC0!._i^^!>;`^&gjvn [ -;#ĨLQ Tr!Y^E e1.gAj.ώۏ7lպ]/Ք:Lm8+sѦBl:q׫Zws@38{ŤЗV6W>}z>zZn5]WR?\}mtukuA=˔䉟#B/3>ds;ɏnQ u} AziM"W7#M6 f"4IBJ" 1U0DaA" ݴrz]xp TՋ~xĖcBiUN7+{{'`"Si5hBoc[͘lQ JNy쩈r2ND~zQi'#xsmϛ\];E?.&"]Mc:ы=ekC[V董"hG$ r^z4>#g"Q81(CGx!p]{tzH%STNW ߌe7>ڟΏi$ ACPVM4AlN(&dx빺ģKN:-x(SLGQֿݤqeu%ad#|s:yGSrgd<n{Q2ײCP|^Qsg_?X6S|GקѧK4۽ޟ5ZHͽؙ><~ -qM~v r_=|qݼ{{{fA0Nnou־7(@黴0nƭw򀞡J.JДo#+8s)zgm ˑ=s%Al8F5.^w>Ёxd.c 1bL~;cx߽?j/卿/۾ko{7Q6};P^_^ InamwWIku~p3OW:rN@-gCpk5=l i||5M=_m24?9,Zsєf#}H~(:uR(9p2Zc&;ڿߞ}2/\/g&{Am IG ?:N_я8p~{w`ef׸4٨1  o l_i&xB;;xr]Tt23e>Y:uktQƦ^ӛ>zeNO`x[r3^:\-{1s8 6ot1?2ldޒ~'uI??*3~'})' X']ϭC4Ct$꫁.:d zR-=> qPI3:L窗$}epF>Gϟۏz2(nPER NC肊##iƒr0dL AlR$ 2Kn]HGE%7 ZXЅ~v/!Ӝ>V L5tg܁vꔦ qdcӑ BxRdc;Ku=Q".41{V,b_ qcΚ E>Gt4oy>STfbT <Û/Qj=_Hk/J3<3T+r(6Ck1!K +%2aŊЈQ, (RQB%5Őnn(C_SQ; 1hA!BT!μ%؀)AX5/FT%y6+%jBN9TTm L>P1QBa(' IEȂiB9֔&rWgMq"&ָ2|!L< 1KJA16V=vfѻ\U8O6 2V:nv0lvq.[Fč[Ra%Au%^0knc]~)EbWU'p.,lc]ʛgbuluΔ{r:Xs c4XW^݇d)V@N9sZG&[񅰞>C. eKxR |oAHr'2&nXS~aKY"T`Me_umq4D`1 b8|hgGa/[O[^+.Y8[Jky ԫBWSUͅ{U㉩_nåmun'{|Q> &( +n !duY ZrCPM3~!ՌmN^~!:-E"wl(2R$p!*8Iǘ*PFZD)JjD wZ]h&87zö %m~7I74ڍn3 1-LJŖQJa(RSso@%#ҡ [m 4[ƐƱ~蟀"In ߻ӳ"gk\|i**u\zFL|g90,M$P$(Ra6Db)") "ĬQ=eSZgM`Ԡb,&32^ץ8rW{J+ gDDP#(h$I #Er5xwήJEXigt@,ӎ=eX9c79EK }U3Rh0yT\e?|xRlp*Ł T?kaٻ榑nW\0s#Ux E`nJK+hpl#9@.%Ŗwɒ{0E-k9ΒeH |K2R6.%];c]$`tcef(.'#RcL1pmuJ.ߊ-!$_PJ1*X>1E^H`ؔ,S>NmJ .;<% RS8"x`pb ,]K̤oJa#,9{$Ь@V6샂v݀1*XM:J3H n/qsξ"aL#!+Jn 2Gо/1a(lXN8 H݄>qko:'zJѯ )R3=`-A /tʀp͙ ]O5#)l(FS5H^b{eb[d|.ۖਁh۩. aؖT3`*9{m!tDr~*ЩU1 p r(emAɤ}΍m/e0ORs7}}Iwi.?aL6u9EuNְx& !$+Ф"O%ٵҸ>h>,{mܽX]o`;IKT'*kڛ5R55;xiھZ9*5G]Bp]GT^RUZ qh^ujy: YG = ~KvK"zy\pM8Ķp_s /B?{MȔeOz-UJK"ܡT#96XƯ]P[Ф0|` ( PÁKP 93%1.WEMHɢM¸&vi׼ye>=D{%y0P{tMcɨD_IIhubuYZ #xj2Cӧ[wU{6Ot:E;poD^ݾg $Gf€&bhőԀ\Bueo4'^XUI nA>(0F&S@ܐA4Ŋb?X8tB)%x- :F!VJ5ۓ f+/a<̨,'֞Oѝ̈́,?Ujwxxz֏qMCPI%B jɇ`hpJJtq09&QɴjK|;K~7q[&ᒋq7 D ۼY "!<\a\27_ҷy9 lz˥Cօ?_()_?(ɐ%R qSANTۯ.Ņ8׃@1ުug?ު{X % rkUqi1gjETy )|jctjl>G1Kv)_gk؏6hy&p48W1n u nZhJn 7vF1El (+3~;ZLzkƛc+}4oʁ4M1aq؏J )n$7ъ\"h=V+'%2rMk?G0_`:' Lw8#j̊c;N۟|su4 0CDž~;Ld*8{.ppa-Hx!X[Ho0;gi&R19~pԀ_g`($J>?o_ߙ\ x0ob70OEOwjoO >Qw7/.}yWibVϲDs=?O|E3 ߊ,-FvKrп;sl zgf=ck/Ww^zV~A<7BKǟX8>ɓ2[t7S2>Z۳q?ޥ G&/2]F_ՃRnn%gQ*V&~7 Ћ(4Ͼ1m4HNEgoq&*T~5G}!jc`(b4oqӫ10d/A ߼3*9Rba 26,ʌ7'Lq$=ܡ sSᕚFy@rc)yRIg^8u|3:.s+zVM&\ӄ9 |336Ue& (1gFE6U!}FasN;h4oIʜ ̶ƪj3)V]л>:55s{%Ty0@ۏ0 ;Pn{9 lG%ف\A룾Y)& )[d?K9nL8ݬ Y+&{=ZmD΢?&fX갨 eȜ] MfUK'f9-7gu#i՟؅Ssmm$;&wyN:z_k/ӗK/zrl'{A#Վ\1w9 ?Uߢ[N,-dɖhY"$I1YฐJz#pjJiba' ծ9̌Kf~ H1bs=bǸخ,NG[ 7)DQ[-5N{9`nzonG{'2+)~ ˝>WO*59'ɨ|E)W[ $JB:sV-V,|;=#w[e^ɫNdě*VE gٰnӟ=iT Ѕ }TiymOts{0@=Q4>%9g|n@&*-*kvםUVA!gJJ+q+}B^[b-`dwjHec!,ݎ);{ \NF#혪q^KKmTˮdmw҄c/@,cri#JvVTc!ׂbV vɴ&Uf ƻHt 5ۂ@lvSKƶdEZܒjuˡ4uQW2Zo߳<# b@鵛3`z9(^xo_xY]O$f\/;3w#dhF0MGV})-k\ ˃Y )AHb03qǎ-k6fRTBD*R&EJIF\"Oo>;/;iȵ5'޵gQ?ϢtJ<zKLB|5Aޏ c \ΰRZ6z&;N>hZA*rW#F6c,V_[AGtCKX@bAmﰀpG8+d0g3BnBY-5T[Z*>+ŝ C̓Y2O.Ŕ%)rcIRܘq9dFh;Fʃgx_{mZO )N̋@ bBwlFٲiwsI[murb[yOZkw4֠NEK-x^4, lKEm-QF'daJ|xoZCxnwhWM n+;#>E[A+=D4@ >VQJIؾ4~-^DJՆ$җRH_K 4%䦹XfM?&X&٦kŖi]$2#qy~peVq^1Ɋ f#kj-MN%gjŔ l/kg{;.J^t|CȃTE~ # J>'I֬Re8w |bsC\[pD$~ Z>FTXsEjz5#ՁKUk=0,s7Ӟ`R MXj7ߋELu P}d3CiL):mjr~21^@-$nɁҁԀE鲢إFΡTs\mÅGWt!p뜘&9Iλn<A;cK.)Nt6pI 9#3-_9':rTqf\Rb0l&&$\E݂N-porgY0-ʇ(6f,2ƞL [qM;;8SR+xNz)gHlJ]bgwj;tap^dM4(΅uct@D/wH+1 l7YSmi5z7FqFɳ"OUH5Dtaq}rU Lȍ#gj,Մȶ{4RdQ*'$Xy> *(S =2#p'rOngʭ:Ӻ߷Fl7xF fhI [%G4r^aKݵ$i \2+knXMf_XYIT*qŘ"i=HXH =22[S;S.Jַs~ R oB1&2H+bZ cQP$ctlc"GZ, X'z³VLOgהYnbn(X^# ,6R P*vF¬Z*CKaT t!j<57놙R|yK4J`T]$kc`Ɗ/()CCB(P8TAF1ÍS! R@2T9e ƱR $"1uզIBjB3-3"\G c(E&kctPtU9SK%a87W}>pcOߚy7X|2?.0\LX_,^')q:s$ H/%[xw)>:]qxZsFV~%vCЕ&|zLHٞ;FںuxѾ!ЅM-*+\hֺ qʀzIܒ&1k%zK-mUI1 Su l=@>uxbcRN<&{^ģ@oIJmeDKsHT]{:QC.k:ZT2zޮp?BTշ6!0F!^zQ2gci24ѱSRϊi2QVFnM2Go,ĺiUEҶjeKfN(Akl':EZ8(;pQhZCz+slK\8Xr>Vy%Ǜq@sz֌jMWz3 Dʍ;Co$"& rxX~V\iQi}\sr _nz2@mJ"OZew'dx7 ijٯel}X $!H!7 }>F`{ȍ9ٵw1?-+an]N nG?\՛|w1-t9yqӔݚ|zzs<4͇IvyhM? zsuHmpp |wO_At0=\k?s_]{קi)7/zӴ]\?\_W~翼}ǟo\ɛt޸w󥶒o/{ýēNd(JAt4nb[coaA/jڝi׭CBwaz=o2T<^1\ހ5NznwƏGaV0?fyW V6}Sh|~]?RA[;g?hY˒tIK8K&K/{_m?΍o{IƌݟOagv.q~$ݽ-\v18 zܻb$^y8'ʾ4' Y5L=_f_& Tvxx7ڔ鷟a$E*` 0wGίv8.^xݍM?q+eTiOoSm 阥Ez{o`2܍'PSyt 6wfmvzlC|Ί 8q=ED8І #8:%zm6:LX`>{ 3pwLjSryZn3^Hϻvg&)?l6$Vl atZMxӹ) &7 S%BhLSaF I|Yϋ3SH(lcA#"͠[C'VD*jS*QȝD(iF5w|۲RrP<ٌ^8ՔH0@ A*BFPgBe#(ƊjR.B6DH(6Ʋ0RcVXQ^Մ0k 6aMPG dÐ9cH,S;gHdL"[3oKWqW 3>oV^伔 QRO jXAv(IG'˓RjΙǁi$TXt!R"jHJ~>2)58!vgVBnE !YA)A4b`-#b Ƶ¥$<#0t*Sd2Hanc ?ϓ*V(B|?05b&Av]?~tQ'5Ӎ4OYY7.l;^O\&ɵ?jr]C73Z_`B 々`\qN l(D؁h1Qd t,b M$1h5`J0tֳ FǸ}{eζ^%Z{K,J=3Ϲ2[!e8 GbqFpkpa`rK >rGB1 H 4YZs0JlsB8MΟ֜*BX*SnYB!_E ,ӒX$քFA*JVQ5Bƹ4QQLs 5~S?M]VsjNp;]Z^m@o_^Mi@Crf/=DME]]zѳ  .U!yLq2Axm, Cg^H.xuٿ0J0P|I&fuF**R'MDi sl`}-ϋZd].8ւp=O2#l3.![{u kY7T4'svqD$a pIbL0Cungm7_$9mKJV/.(J_F OV"3+i,:qj\^VDZSZ!vZ5KflP^Z!H^NR3?M,T'So&@ 8+V5j0%jtZ P` !(*,6q f M#pjoElZ\1m$|K"'J%UN{mΩNH3ǹlnݮ8GrKa2+]@hfqH0qT"l5w\!a,52d"Z Gs K9pD9F[5hwڡTdB|{`0RBKbLv,AS%T)^A w=cejd5!.%*9x\&P==[xn Q%FR1J/KsC$hq[<pBx𮶉-rb΍%>㰅pYnb`1sJ g9\ 9s6o[ ee8ˇvJr5Sj+t>EP=@?1M1H2HHk0y[Fq#vCRb߁KVr)b էn^a`pOaNe 68r.DU<2a53 q-KܩցH+ Q:hsrX=P>QeBS4QhY)qa f!\q9m <+Ԥ"g(\b^q)PL5  WQqR\?j =Zcңci/=t) /+z W"ÖRjk2[(?r!bԃu"u# FLRjl JSܝGNȭvms'!o Z0\5-E;erI3HwV!MXFHN˓n`(: KTM$'64.zhQHVnڮ84|)шvI;hE\v<;rvb I6|XHV{!o&M0Ո0=/- ҆4V3 ,Z6!~s!cKP7+3mc{{}P`ߦd|h Uv' RmzB % v/T $Mµ[UX=Ur7PNyyns; >?xfnҹF3o经}ytz6_TWbnFnsF+uxƀ-8R /1I4:Չm&҅; g r!7*u/ m$]\v xhHMqE'_O-빟On Dg27թǿd) mnRz˯]xO7+qzɂ*YM@CPaAKkzT'VoTgU!UKNOÎ^}XNF>#VVi GI9U7Tw:u)\ӻ޳GA%q$ܒ 8g5\,[5F CX0)84bN^oO1I(HyzLG(6zN8xUvwo |X1;ޥ'h;0T06BK]wf"&51!)XV ͮ4h2;aSq9FNygSS@V[$Q;rl'iXXMgGW!|5|u3d6_rڵijvM^$'x`26tq 9h_v-NGfozVQ٪-He6L ^F#M*l 픷9n&8,+ifQ GcSdC| 3&xƹ5z9EPk =:*zȧI0rF|iK 򊘟p)|Zz|xN)rSNWso7l,?Nfx8U=LDKG4=w H;|+-+L3#H_?~<wn]5J0.?X:Q;jD ;37kpt?g/'oϭ9dm![+w[C1r? Y.g)ߢ1ɘ֯6 +̋8/x UF@k3lzai.7~3(:>S܀pIq5BK qt <J0ޣaNF}M݊D)13JnpޣwWKĉ)J#6Ȗ}7=V»YlvEA:$pE!*1ZDu%{j2:ڵJRFVWodODWe8p;eo7 #m$eUwP;J;pRQz8!:\w9aCHO\(YJH\8=bLX5T`?P>A e/GPBih;P_D f eK}QmK|'4m[r Pj\T('D;th-gLmu#iWS|˶p ܜam՞dd\U*a )ru_JX9h(;]a8٬(3:~YDCڶT_ð]e&|Ax4s G8[޼41#{g;5XxLOT Ngб) a3RBj {~"=iބ ADQlk4VZd?|,pQ~<(ѕ A0fd!b3EΈ0FIJ] FST'@xV1`Kn-߈A`g%q q׎#ߓؑR\<"h:w9U} an8Ͽ܅u$TB3Pj0Ԉ":V$X)Q6 H8)e=f^Ontn dV1hfqq@*wq| ,cYUn^fg^lt ͹A9P~hpv @릢W\kߐ??Y>-w/Gx(/q_[ a9t->厹.ͮʯ_|VzWrqLc K$+1hMhѡoC!!#Avo˃v'g׏Kev4tL'w('b:tt=(/~_\֌r S{gf9sJULQ>$7k,wFгRrN3_gxcyf'x_3T(Q9i)8Ks3L"^w[~M19t@\ ]~l 0>C;mJGVVMDpbDtyNiKxl 2ռc~]4!RL\uEeA~Ab]m0?9__+&5ؿfj{8_!e7rT]qrЯ6TU5" C%ꪧ<\7__֔$\}F5 y9 W釁)rUb!%=- GXτnc%(GK4I*<{ Bd TH>Iuܙ[ 0t)N./It#NAa?9 5{\ap O5\H=wLhVj xJV B&rc-$J&>oJ9J|3\??ӷd0W7yLt^otl'Caͫ\X*ּʅ5ʅ5'62 BsbHi #W sZHsraE.CK}ݑxo!R*okR\礥)\G7{2AxiUJP]"`jzE8-⤠L Ky"^g߳ ,:m_|Gn*'ՇDa-Y I6"dMTI$@eʫ(Ձ< e#'h 7;dm-mݾ(uιbڱEp\为$hӦk|E}}-ԄFnYҾ;J-_R+kMN욈F) HL3̾C#AI){51S#Agc ՗% i9*L?!Ab/1_`(4{caS. $2}k6PUhߝt6Gwz5{W׽܄h\]8VDmH m*$ V%.%]ю`KZ$:/G78ZL668Y8:ٵb5ͩ OhY.k)NU7JV֙Z0 ʈi>g|6eڼX|Z9)՚?i5yH$%}Ip؎wE")J%wd[cJGIlf$@!aIS|@"8[OМ fj]Oq/|9Fl$-|_Omepjiw1ija!/N-م>v}RTk9l8uj ;:>'r yD:pխ6D9vosBj>%:m( J췜}muyޏ- Jw0[sJngEv{"w;+tkleRINJ6j! kUpX$K`#UVZ?Ɂe-xWPZkYAcy6J}vʨ^\l5iz L5eD{F2)U:5K"r"iG4z3rQkADr_h~aR0ѵI7/@ܧծr:o+sv(Zty.@6mzkg_]Nތo7u~]'/\ M Iz-w& &3 0OI`aV(N'V`, pB1_;tņ1LqwLr$a3j1{# Qr) x?JPl|:Foww8Y6=kqU:6q9~ty8~|+9Xo;'w(':xP Djյ3у\ m*٭qqlf<߄|HC}@O<9*9JRٳ5Xdb|q/-ECOX8 yzTN0-妀1̣D\d'Es"ȨzjsX"R=zEa5XaƳ|9xي\c=rmaAx|1j~9IΠE}V(HAIAZ:#P&ySz΍P[0':˭ )$^O+1NӢM}9t.y|wstzAn>wW貤ww@: zgu =I4WO%8lG V]r:{d\uo]w4!/dӏ/i8h8\ 1Bd\ }H\֔Ng/߇ ׎bR)9sߕ'V>#ͪ\Tbj/eh6A "&]f0UrrNQFVPr*N=p)-?BߩTgS_q,6"TzƊpBoTύVͬ:ݺ0|r 0o .Tmzr5 Q3 #Syj;3JBD^̼H<=|r K_C_=o]'SienWuFkB}F;obg%/B`dGR,R}1X%HTa2%Q$ZΒdyJh Q8qw7p&T$\ί{=9t^[GvcbwY1dlGiEʅ(S5cyaB8G > ?G;;L!UJ\mc尔bRWbpyש]'oIn  #%'t 7;ggNK)H8E8X \:=J#4J9$B`-t4^9n#*+vg$TiI?KtL;m6]8/B&ל@Y$T!}kM"XWlqQH !NgTX(FFO:мf@ EW9xxMf`K׋!H+t.KjdtM*αqc1^qӊsꨁ<&"װR5i痜*`X@өtt}Ε; 5VZFA@L4yp8Bw \*d :=#'V`>6ޕ6$"^՝W@Xt z)ѣ$$ExY:HJ|MJfE} %RN$I!qjqHHH([4IJeDDUW"r<^佧ZhDH8%",UPN{(JCB< n,ckBb~GD₸4ヨ<)ǒZD$8gFԟ"ZF'#QԵlAtL1Ḃ([zoE!Im%$9B 9 Qc}Whl<09GC^B1";MT&G mMp:u y tm>s"ONvnB WU"7g i_0Z7a "X7˝`HZTR.awEe媉I#6 55&ML|\Ƃ5&W$k)E \*\H5@r6W{PM-9/V3'`xR7gnR7_-7#ZS#ZrE߉[+#Y3G99!X]V5FB Y4SsIpHO͖"we`hDO̭&*4?S)'2pV2%]m2~"_JKKJӸDThz}a/p |hG5ҟ3\0% [ ptĬe(^)ǣK"x2]V˼V&3;{hINqa4Xthxo3{/WHHNa5*= UհB {* .X@VkWY~Ȼ\T~iֵ,=rUJO#sSk,5]75k(+b[/ (mk~&\Q*E3u(]>\h 16Jb߰W ڇzo|1 g> γ1I{Hb:FwucY*u bc "w}JqKnnuwG.,ƛG?5:!>5:5f2<~FewU!afMQvÞ}?#R\&ڭ+#Ov= V8/qyd\ޠ+ٸ2lp<3#fޤ&f 3P=͇W2σ33)F6vl5Fs>~z@ |D+;k솣Ô^__Nac0=U 7!!}^1ɿ_YW'Z\z鳰Z!V S?Q8y~$>?׿|9EZm4xHr~c܉X?TǝBu>YoY .UJ0S) N3+(nAShRM:)0Q @ iNz2XIGq=> d[+ΐU.OheIxmTΟ>(QHt JpMhD"6& [l+k֚4 ls^Ho)QJs`4]6Lg gU1NQ欪zYZ9^S._ 0Ju2ϤXkd;B (Ьӓ 4)G3P a3sV5; I*\O?z(2s7e s'\rcz&+<.S'x9Ʒ!<441l iV!\tkģ](_OŃ^zucYQs,eJ5xCcV,?LČG y:c/$?b#Z)ʸaZ)j:V-q gD9é'aUlyx6&_c)]J=mPOιMD EEr#kGGdt߀\cZ^ݖx3r C/J LbnzPUzHy͠RF(ڇmı>+^ߞB7?vc2ap(P0DUQ-?-O9L֎Ӈ hc Mn~Z#V2o^%޿NM1E6'KAW*CP+ӳ[0pqkW7@WjKVtG||hJgT{*=7=Սx҅nf;.8c :JA-x&,Z OcΖ317ZZfjȞώ~ EB>ex#Ov.FwK0y9rZ`ڼO$hrQ^+GtL |A rFbĬ^)&毴OlO }0g?CZA`Ar HWw,ohU!-n ur;òWB% ƻ$PwH2Zd?_ I%4c@ T@Ed5DXXL;js ^IC_ZR3Ө 2E{NZD2q|GF(nD !3A%Gq9OI`hMp:@#0A:@35IFAu{̟Q2 4CT'a4.?5.P Y|C5=&+8(ґ900NRת~J'[Wnr} +L4K 񌢞wM=Wp([XJQzݛu[t_g\xWuNUMGŀmqN ˵oL-VT9 6ADkaC@ş> Jk'FAD'۞V' ugREP]A9j?VV&ݠ;K'"&m╄h,\E#+;nî,kYXF>-k`g8l^6ޒo]͵t ̫:*'$1qZ)5=ufO#;Ū{p &c(:z3NvkZП Vdat7EpM?^OVwߪ=1l>Š6`TM*>iNޚC$a.:./pیȮ˙o Z >n#&.$&qpNk Re9 +y}k: 8G1T,[Q cxMk*8SN5c4yaM Gw1b1cMS;~Q5#_@Z}-w\!F+ƃYq#W!@7 N1͹X_ o;:$=`*eU|[F ^n뵙3B+5w2 Z5TeMMϧg_ e* eCpƻb)SO/%s"1OtH Q_0ALQKzϗ=h4ÿ|Ż>\azLxM=8j+7/K@$?Lsxq&#?}@ _Eݍ&q>J)bB@a&VR2AJ ܛHҒ'- 5N1i  hTETD Y^ ?CH@ DfH@׻׺|%5-$G",XT2A%tHGqQL.3k:;v#|N)+7OQ@ 뼒Y9\|55utO&3Tg+}I䄖{O[J.{^O9ݥ l-׀,iT J*jΥtIH(+R"%6xR.:#8ZQ^ USlk+fu QR4yAih%*[Q@2Ri1C^Z L.n}Yzu0C!,H?LH1(s(mҤ\J^{YqE)@<"&( 8*z Juz녊G+mכA Iz~{k4# B6ș51;$O|ɿ_$;K-w_-oW6c?!3 L/ l!1`bKZ[@0eǯuU{#끆?ƣu@^?[=Sp&>Wi7WW],/_Dhondƃ,w?&wxb(@^bLYК]߾ۻL)}Z Y%}cj龕VdSN'~~(=5}kH$m[Q&qvdH`I) #0,@ !uHZJJ30k^6XOU\no̦ӂ P! C88 #c<ӦQ9Mk"rtkO+~s>C1j I%\m$\zlOnds;n{BxrӊȭkhhQ(yiM9zǏ+j p"*sIE=D)Dj YYy [,y7ӷG!4dO}'p9.{cɭx282W:xq3[~~d,{vi hڧoZڴZ@-ƫD&\£;ܺs*c]0EEB\`%;'if)֬vk=5&٦udKdJ*)VQbF@I՞aN6qAŞQc9qo7glDOfGT ?<4|@^r$x"q83\#DZ ҙPU6gV]:8^(}~=m!XjV| YD+BAZ u!tIƘtHIT6r_np",1Lo|E8:NA! B­@Ƿ"B䠵=ZIF D!}0A+ T(ZPX9^s`4>sCixg{Z G#ɘVQr9@.hh<Ty%%TUª㭹37KVas?83]9lqdž5qtպ Qav|6,'g6݂z EQsY@fjC@EArHsiezDsФI+)='w(n l] ?fivAzAKbBM`ϼĀ|!/pD٬-1|T30vO!0n wktz[:qG~0Y*>{(&g~OD#cP"P1-5ulYs,*֠~th}Շu(Ai/T%tc*_DANn[A|"pd7~:dc̜\1]3 b3WP9[BǬ ."tc,g&v&~V~=8ɍ ܴ߲,'HkZ/`[%!P⯫r c=KV鍟WHmC^8ff65kim_{- htWVRkZ=w杻1hs2vճ՛΢YH\ RX'w:ҭCq"' [BS[YtRQ;(I#C_hgC#iuHȉIe(1U3Ǭ&!)ꆏy3xLj%6R5(ڦAb`)j/'OsWV"Ӈw]M+Bq]F$mX:Si4a Rf0bl,3C2n:)8P(cnhcUɝ1S nf8Rbx 'V1`e uPr){ IASԫf'?\gzVHQ (g nbaA42 iaOBm Bɡ.B7S;f&sy"ArY Q0\ g"$L:PPY/"PTZb哠r0aR -[~cItu!숳_֪-X3H?8 K^jR&{12KCQ설E C SFAI ('`< zլ4pݚٿbA~|7Y CXCzX(Ƈ"^8^$86BЀvi3&2CPWJ*[(]ZcJipmFf8Ǎu?m> lTizB$蛺62Fi<9m]Ɯ%Dr$-"^L`J;3lˠ\YLa*H*>r-0aT`[L&T0w6dp[8PuaTP 4D9_V)Go9 2d`05[bb| ?.CV8c` V)[mLkQ@P%8^{-A(ÇZ qL$(r0#@q[簮Vk F8 4Z ^ K78Tr Wtŧ5UrOh}\Jq}=&V6L??GY5QWg3W&)j5Uեbz3kWOIߪBYtP *O\ XPC@M3 )xnFqk%LW+(!D},]^9pDQHkA)j>X@r~nszsv'j /X Y!2Xku! h`0)9ihg[(f$ LJ'5ܨ5 6p v)acP9b/)V^ puUdS*> })BRZLm=):;Љ/&Z9GL7dK#✯3g|Yvo1@~|烟/bxٛ+. /,^S>y3-`- s{?ش|0l];o#32~CȒϐ #$QBG#ZFa4q8yLAD&`(VQyRC2)_Ѥ5jP5D'lh49*4 ڽ|*;*E1v@(sT,pb n8GnO N~$gșA>f܎9{=.C][D6pGn5ZM;[Mj|ԑE0[ۄ>Jí;i/i< P_GПޚ̗;LO2j [qa .yκDI{E'lsYg5< <&%yL& dHst~~:'J iZ5֮j B T%B|9"&^YS7xG^Υ :Rl>5f0$Tрժp:eR" Jfd,y`[<Onfuْp]vg@@ݪUo~1 !#bxH/ָ I<|rے@^cg_؞ zWfZLj|,G}\/QSiBuW‰ܝCv` @BdTZ <`9O!h4C`Q8J,bHUc!$~x=*l,xyAPP { Q *)fbٛ /@)dDvnsYnlݹ8 W*ZH-ǛxN? Jt,bw oF=Wt,YhyZqB-Vwe,&Gڹ|wx<`rWYh^\ /X~h5(z]E$qP+ְ\F^ G4?\j^D}39!3숦Se*Z0S_/U5nAwYTr@9;B/'7Fԧڻ}CoBegq> 6+ -Di 2w]?fvٸgGJDs*%:f=*/3[ FFиdDqT$JEz ?{<j:+A.σj{E3߅w"X%TKtekh:vC4J(]oᾬӽx)ʅw'Ͽ_xSo(M;8 ڣj#nPl@羾@Cd2k-P#&Nb*g$8KS5)(=IXha1(8VlokYg%ɇ&2¦%DPOYsN珃io!m&1Q,X9{brɕ^-R l>߇_VTj)2тX s)z2j1r,Dܹu55M[IZB\ؙEDB9ALuH{d,^THTDi3& VpHl$F%&v@(1qliB-I33Qes{O%KH@Ed} 222j ?~Ȍ0L9H3F ojI*)[+f}UYToL+q lWUemz@-0A@Z{.v֞UX7V +kNᢝy]q9|WN58 R6h)&b')l8N}wG uV1g!,Z|~HAEKEʂl:p_В.:BPVA?~&&- 8joEYTVo'~#uo8j%o/ +:9wCW >= JgAV QKrՒ9،=4yhLPȘniXL[zѭVk[g |uӢյgKKDhEXa syAZuK4PSCVFᨡF5Ӥ,.CBK]>.9}8}#lU>k[LS'3#0J*c2*"|%cMF2*c1(\ۖNagʷU&-+sWbF+MN&%Pl-R9)2(6.Wss0\,Kq(Rfi,!Y QSHt€%f0h#8BƂFR܍hj"C,u@en5vy`_`}dhj% FNkpU*QX[cc)8K4לT@ZSD:_,kf4V8O$ 1J4JS &1"|AZDV ,hKCcJ AR8W_HTDgC_+*ЍI9SnTO T YBlZŝpi1*I 8\&[8`8eUsԗMv(REy&bKu\YYRSE:5\OnIĉI$2y;-&Zk)&#ܷ\q+)IBM >o(~V6j~ݟh 7h쒜͓+nSZ)8p(kԚ 0PYEڥE+K8?5*ZauZm@J57$e1^{2T8ef6brAIPƙ&^u 4$;3|-"AbClĤua-ſ&q 2 ]e+XH 1p.#յsЪ>qП|+UX v)hVu4 9MόXA]ʵ[dD!RED7DVQKtCʉL uV"ԆPlm`L&yjQdu1(i) MvVj̈́[iB6o4OB#7j~y~wwذ]N8 ea1]X]qb cW.#9h ӣKOmCSs..d&s'{wtO)&Pn'4&Ivپi<Zp8GC·Q[vf~x p$?uy Z2*SL5OvPۚy4$juB{Fh| ^3K_^) !BeNdėD  } \˜=Ҩ2'Fֈo*f]'~kV?0E݀lDO`ܠh> %O'(2"zzkEPDh]+-<EpAќBkq\2 Z N@4&fPOwpzO"xr4->Oar!=r S Wm)1-ޥiGiAeF6\ cu#F+'lK+ДVeFbyVu.!] @?!=珶η54BŔi[dpӭr=5:Gc ޚ!_r+)^&Y+Z{Vpu\{زͺe[y{ a,h(QkiL 葚f!6A%-̩aj#K_zBF5qBI&.F d2)Y;R;rdٯ>E0 TFɑerMtxĚHR)ޘf\IbH3i4 hl}2' ߘovC7B *L-{ܖ$gǸ3媧5KS"`YIRpf<zm~wݠnSTRrT [8NWLMՆEZp!a-wu]P.kvX!(yZ1&(#TZr ݺ{w"CCk<|j˄֧פUD ׀o/Sh SއRC+QoQ#9I ѷIR3[YꐴPޯL1Lx!q`~vNM3;svvҹFI>zOzj^ ]r'=ܣs/"gg+pH?)!AE> .Y-YDO F-KǓѷhZA{G fgV[|It+It%!v!AK{JΤlBE GXȵ1b?Q?OFwGn"]s%}qGf&2Z(R&LeX 9kRQ"خ̂:ٰY/6ygF /^/$S\HVCOuAV@Cb%/G  bņwSpE=(esPKIE8 m-sݜy1_lE٘fMCsag_jc ]8 k15c"shw:E5o\PCk %ŭImx_"еҜsk6KΑ %|ab/IuJp!QSӿlކuSP' 5] v=dMm!p#%U ĭZkX!}l'hrX}Q{_Sic46_ݷ9?TShE{/bχé 6uk(ۭqT\olJ>z;lr ޖ _]%& I$L_,L)YٗX9)%ZHsy+Cp2+a뮤eKBUS'eB,&mIKX#'@>*F WX \Np~xvJ(9;N6ʵ==BRqPr* 0d{Ɨ*PH|xHH}:+1TsJ ]dc4#oqyLnlB _WD?f oB<DL|]Wk\@g<&Щl%1e18a4̓\d޲3iꗊ'%.|4qA?{Ŧ;j gqAGYGI:3Dt;oЧ&Hs1ŁFn&&7߽˫WOX/|jO7~͏WnPU^K_+nw}z=1h.QQ2&|6-Ru8N7lBs0tMBS0" ͵SʘLN+:ߔW$;\Opyn쏺ϷJٻw:yDpNWT,K( ™dٍƾy?{ȍEvi,@;ٗlĖt|¨ᱛ5wai@CjxBŇ ӧFTxM`pG}3;/@GU!ۂ̧wˢҫOKw(*;E*)Uh>@P[2uƖ-0hA[ jVam՞;A5}?ORcԘ1 ǫqQ9=Vwfޤ$^g@ٱqʹ ̸u>b%j 4%S?.@}#syP/}i|N5^bSjnڑce-I̥p?ƅIJq<.fvKōyp3 \v0AG]qT78T7Ღ<ňs|-O#.5j{e6C*]|'b~F u˻T;gō!ɸP}ЮqsΣm P ,Z#;yǰj ]=9+kr4rq8qkkCh@!y' b"cYEZԏFLWuh`]be<.3g_^޵Lj󻁹D!ҧ}`_Mv|P񆑈]^cL.RZ3YmNBiLrcU2+,A{\c1ހ u$&"!R^ J6^J戥v{O^P&Zg2QJ*IT(i8w"^jO4[DҞU@>GI6ZTQ\1=}ːB`3g, ,V)S bFfzau2'3F*aYYY^c晖Dq!80e&vؓ(N%(jn 6nM 1bT6FxXDPM=PA- $F}soG&Cpr.}*wp]2%~01H$^YZ-!g;Na kikHNyd6pBRAJևJ{e2B`Py^GF: 3J5d;n~*ճ^䜨 C>b.)ڵ{֝ΊhQl,Rت&~qWQ8sz,jC3 a ,@sWi'fa[iUX([mpE}`tOFS *pT}&)gkB:ޣ"Rav:0,` AN]V!ҭsZ !!Luӭ4#íZiƌPJBǝ^|'BTQ[3'O2Ij8i(a,)x ,yZԎI4aFbJAuRM&)ܥ[1aJ2}T 01: !U`p˔6`0=KIyU>3Lzj(^1ϡDs[4_x>KD}]=ZRݢskU>tㄣZ QN8&IK@Z)&3 b`}:;Ѥ 9cÍÍ{|'[rV FGOB q=WcGK2U(Cfg1b%~}`fwzGG7L'|2{ Q m\VQlQYitA ?_|1Ļy NGnCD$&/ζ ^jg|%?ݍ'7`Ѐ,FY|?sgd'IӣÌk,9DJ}5lrwS8Ta)v*.Bm= 2!)yu+Fup$8a9Kp.˺fG|zk6~$㺆,hU&{Uly6uuVgx5'NsOyoܱ_'O ;dVqJWeW934G҃CRW ,qJῤkϫ ł훨-`DÚd&^뫺Qش3 AA@7̚¸VQaTFnC6'UW{5(.@4H\`5]pkL1SEj *p7+.!ךN Žqxjt˜2&)#mtpJPe)A$r)9s4"\BF8#GCH ?^*: d v%s܈[j+MBɭ7w헮@%iDU2CҗBsBhϹQ:ZPՉ[=P9#d3 i 1OȼR߈ⅱD`gJrz5wԿEu \4 7SP; V?@Xzdd"%`Xb1kΨST0bSa56bZܨUO/Ժ&\Kx{2dWLnL3,:"9CΤRLY#C8Ұ VzO٦FRM P Oc>N,ZY'Qh[6WbMH]RJd*fbǞB hSmeVW9^gTq~ ^3ޓXJK$/b 8y*1JRFȥc 6snT(r,M]6 B܏cD),!t Q=nEKTh_ g W^d~o NngP㶗AO']i+9HP"4'gj5|B33R>ΟȞהq|BI*81<*yǥh^q{pѺqQ _UЩ~lrwZC4ZhEo@n uy3`S^Nt?D!c1}5Se4-'\ $|Tq >LC/Jk^p\&F lf"u}X1g_ V_;G_~2f![^j{*Hnu}dr9kN ^U&p6DjF+1 N9AR̸PaaCBj B VX%&Pq{ s<(* ڈå=h pjs,0/jõהhM,kKAmd@JR=" Ku%".E x27YjU %͵1 VHPR8\{DL-=E~QxOBT rh$ht~2?znzb9c8 HdzbVq3: B@.oF5ˑvPOM|3 wrѴZأm |7履'-[2l?[WG%qs=dMGףw<oG>N=“sTSomM3~ް*͢)ZYNt3./hSZ== C Ԍ`ս{|>/6&-ҨfZa +dl)էVȼ=Kk6bIs,ir P8ξB9Õ,&ݟ0%n!}HԠ\C5d){'K%R Am,_uM]|VThˤ><ڻG(6dlSl*<Ih@I'#φtjmR  .Z Wzˮ NwC?^jp6U]5jI]fT-*e6fU±!!8f顔Ą$&nnTRg:qD_FD IP5Eeb4FkI()RRThBgCHT5zנW UjRkk\[p H2dm  pehQ ig%JWoI'8ʆ/L}N}x4_e2!W=BaVu%Å߿p 3Zl>~Wp?i({9vl_ 5lX+tu;qgp[)gj]kҪyZc-J3o=zT] UApPi-e-UZzoAjq'FÙ`6ip*z־CjD5LI TTk@{ dKVSakȼ׳qBz9˧6cykswjIV)U$n>U|gCw<9%4TN8FCGN`I"8_NwW ]yZo׳ b_9qtu+D^)ŗk_Q\(U:]F.Uݠ@!4Z5Qi\ȁxvzJъFޛZ6ԓhXu PY߽1֣5[>lCydF4\-~pKcL ~/wW7_f:q7W%S2W &ڨ{gijN :*I"{PYle pcʙh}eĆF"̡uZQ_=xt2e .`P>C@N!QKxS41cW-6]1T*k).;wŽsA]\1h|[C}d:)}6X0Y4h0FJPC=4B`'*IVC#ozhߡ׹;9H&R)TvibΎ8M 3AeI$J gxvz A5ty,&<&t"Ssm?9@6 &Tohڒlh"C'N/^c=t'!7yp1T$DF( Ckwy% iy4t`bBƄօ&`4M#`Rth&ٞ7 %JR e.>J)sfIU mΆ@%Qh MYk7ZcY|ư//Q"ڹ(G6NCF 0(Ǒ{tc8F x1z",yodt{sa Zr8=ak7 lBfa/IL_f&X+PH;ɱf˂E `@YcBB@QA挘֐ ]~&v>R*1U^T2@)vㇵk!ͨ 駯(`#+|01uqAm;ѕߋ}JHxǐ!gLv zWrXVWzz]|{+YV:UwsI.<;vVna}$ּD_qbpPu I*rXՇ:RkN{I?; sj^L/^f9߽^oV}b.>lgEG_&8^r`Qxsd2تo #6_ރ]ޒv%^3Ω+`)ihHڂt='2ww6&5/ƷNqΚirJu[N0;4E0(I8Ft*0;ό/J#,Ss9y0k1Q̠o"yѦ;,~F>#QJϳּEdKAH42~ULM)攪HA59PKBlA'=pOk $DZZ)f9(hxD ۨ߀> fT4 d@p6kl9[´tSżS TgmfS*YA**CSS0LATzAyCBONKFED'jS v]ҍ[\bgq(1Wjf!cj-U௳ů!@>CerU3NO՛Tw5,/P>E_qC)և 1#Kb3>[MP'u*F*=|*xku)^W=V9ɨzg_cvǁSjc$^>W//n#z䢿 {gq2ZJQA io<;#Ԛ [Ȧ듅d`!L1GՌ3{\^:%48\`:#,;O>?>-U#m҈-raR)3RL*7riėҡk|9CVҕm 7~dM5bM5WaEi_jU4iĆRF+>,Dh/&T\v+9# rHTS*g|=фK&ͷrzPԖ e0pdːbCQfJڛ̠ʹ8 -]Om_jeՎǢ+E^Z+4Y5g"o2ax V)C L@wo42C˟9ReuKO_u(͞FO.g挷]ݠ-k"7wNM'1L'n^n@)D=g?}Nܥ2],λj:ڀj*!?$4{WY.+H4r9&ʢN{cZx0rVQeFG0tH+h\ RqgBTBI_ū+` ~m -q6ZUc{ejn ZE}shp"+ _w_ClWY.D%.x!egw>te9yX;"LKVX[4)PïVOenm -/ΧR MI-I䒴UZ=RZLnϧ7 rwi*Eh;3,ڏlٻm$WT l٪Lnj벻M"AVFILf+ɖ)KAHI*3#kfwx:嬱 ŪGU85к0g#z Z;;*\) pkY V!߭VB 7e RUj"O81{X/m0L]/O=f3TZ=lpmaX>Yیּֈȩl멜o'n+ɾoz2<{'f˦Ґa5jpº\m&d|]45$}[ӪPSF$ar&ةq:;{ f,4ayޞ~tZ3IY=oAs(Zi|XOBǾ s1=]bGF_?6J¸?I2/$3Iq^i\[WT]cMk.VŲ=qLPKpmhgZg8~H_Ubˌ(ȸ5iHUNq[G=ѻUAT]w;Ĭ)'n_4Իա!=WJ:itzRU{ !DVbU(Bj8.:Y>,0ԡ!=WJ:OvpV!Su2|L3Cnz:4*ZI\+SJauZssy4g1ahi˜kp|˜0~8fkj*P' uB}hVrsf8s~KFDzw%mF9f+t9αU;хiu9m͕ķXz +l0>7Unh %Ɍ<zM _gn39 4H*{NpR CM߷6Z";@Fd2J~~0.A"#4W 7~<)ðnBTSoAR&*FTg 3cԁ<ڏ']P4auj'Pi kHct+M$n`/GM]M&HRV"E *R(({ʤ!M#qN:6n3KS)6(JEEJ!9t2g|in)ŸgEɢ{ 4-Ka w`YYفlp; DR P$58 \'I*fNxjȽ|4;x/- )yA хټT:)P3$97:ܐQGٙeGeK[ybgt%xn@AY"@\:uPYI6fկdܑaO퇯gt6Q:QD2d1'T H9Eh T'Y#~]kR9^}ƅ8&(.0ʼn/0n- /7}+@e,Ĵ?3죝.;t3ƢĦdޫ[wY_w26W8ޏ߿xǫ^WGΏ5Gam;XBTq:uAzfbЭF˞6Q][~T(&жָ '2, m!kz;1Pz[C;@:>ր%kT气g7Y?{{T6q[un_1{@5e-޿'#fi?{yj]6(#e4}|1Ӵa9 `i0Ղg; ep%+^fջ(q,!GTq)$+QZei$eSm^B2㰞]m+ApUeV ?*%q3JBE*o)T8c8Ǘ:7N`xr|wA PBh4k^֤ׯ|f W"?MJ_\(("ܸ=A)s~3QϹP32IP]TWaZf{7[<͙継XXr= NJjOˏ Qm"͹'q$rZ\1Nsa/f1hm.|~;2,Ǔgښ۸_ae/gG*?leT'8>M sE*$%{j6x/΅Ø)ǶLp_n_xxk+L`~Pk,m UPE-VG71lj}Dd˼ m}BC-N )u —s=حƜFWxޤ91*Y╹p6rӺi?OHtI>ܩ m]neg5XFzk8Jt ^cfFb$B"ŤB28=6ZuG= ?H,'96]nd8`|0W?ѳz# M+d 0uLs C{dbs, .w=eҕAli'ٽkkTvHb˕(&H x1<"ynVH\U$VPZhl |{ʻɃZLpNBk(uB0ۘBü"Qƺ,oGE<(g[X1qgRh{ҩT]ޢYy"2`.M7ZTK-^1z*Ѭ{kyœ l{DϦLN$CU)+{{Uf7 5@i_-Ƙt*Ż:iڐ=M]4Q>$~xGY.TI%/ nQeq;3x$!\h{wSXS:(Wz9![$4tC9cHArtR0R1Bx^gsKFڃ¶:q A}Ixǟ*9OO\EEH;V6>zJUr"=-hFOPltP3^ˆ= ѭcZ6#vkQPذjߋPo.ϙ1 aٌo}{Hy&7gxcX`&}cE%~adNkʩTPIsX#k3@ݴHfٳ$!4ߞVo7Wv_^;/0d r$TIiIQx0,㺛v2+v[}74B7?~=|1oD)YƷ[ Q>Iwxǔ ;S+Io9/Ҝt=bd 4Bլ@/g(W6wlCM{{?{U}(N yy&!y 0U퓶^P~tz&s;ݎ#[c`(0`d A rL3Dō0$%SQ`y,ǨvZ7sE3IuQlNgKK^%upk_k'yKQzU zvQ3Pќ3wDDo3B׺8QwgvQ9]AgmZHuF7AU!4Ƶz|X:ZdʩksLjpsáP^8ᐐN“Í;}a'8`1x^ޝF'@Ks[TLJ9|Ŕ48U=8zԊRđ~K2j 2Qŝ-?o]w._6BXʻt[ کfKWg;n΢.}r ,W(t,Et#jz&b#ZŎ?O[r;bLPJOZjXd&2M"J싒.[Ar&C+q ZOmR7=̐UMc2Gԯ[Mj9TfN,^5ʬ(OtuK9-^{ϰ-}/__I]_=~e}?ڮWc7WP*.5s˥H;00X{Bl8vL]_aR',x'0z/T_9M.4 BF` y"# '2>Ko5 j-}HN'gm?0!LL z?Lǧdnw51W_= f]c WL5@60Ec,N#Wт?= 8mϋfMYϣ3E)/f>n{,w9.a}3:/{HkcKT}m&)5sk ycKm>@zY m3ŽЬ 틝۟oz؁sMl Rb01"A"2ќ3n bw*h̥ŲJE PX&j859pe7r93f=zӛg3"qbrS~7ZB@R3e}s4:By0Pb%Ι;*ryX#k3E Q[- au08Pxs`d;JYX-噤:" X_eXc,o x3A⺟Y럷x[ԛL\#'N3'SݟdwJo[nZCr +w'0F @I7<-KeIJUyNE,ma'/Կ^,yˢw~}@ {"yQ Kj9u 駹{y,|Dn7[!2 h#ƓC&ej#)0d1dҀʸP&ZCV[=ZT.[Zo Z Ls>5wh+FЛk`kjP4j[5qѲNVLu|uFYfSajKFUV݇V5s{\AU+ ,\@dMƚC.,o(<_CZ!E4s&6|/ 3uY,Wź*8TU,A[EX)xe)UrH`rJ] :[PA'6:˯fY#&Fhy@OGt3K:KR9;SCCjMsٶΩ]ZB-T{~܅;.UB |S~DCќxI3Iō]L[#Ne"9`bNHPT(S]ށ\jAo{DjqMqpܛtzUW;@,禮dDfsӻdX- ? ^,?5'>3fvjcZ|5#ᅯ7cs9 `? ?/$2љج ˉNV:ŨNrUd/cK 8L!1䜓)ħjYJOTlX$~YLM߬˒F$=PŹ8ʟX#^Ic X*. w)o#2!h3K8F%DŜ'G@];*kq5FDM%<]G!qd,=cE@Y;Y\.uoJh<,ϻn̓|sqvRrCL)of'/.3O~Z&? sLVxQ'48e0 Ó&yӪ<4)ocfjHM~f opܲIt7ozԴpq0Q.ى,V?owӏͱ(Dᗟ^dŇ!S.+ʚ} St`eN,sS3B N">OZXa&Uq&ДKκ53٘ST :9ł(qf@]qDqkRg:i?}.ۍ3l+ӛmYN53>%~\ME0=P[:|Хx'2 kN 5ꚝ(w\v'ϙ]VO MO MO MO c=5^OÜk:Z!ilen`UfWeSrZQς9S&jœΨ&3֢ ʲ7QN]hDroAݴi/SYǶY %pAU)nF$J! l*&z("֢/U^B<jEcr֘xmz;<y}4b ޲ה̚ B>,mChFV'PV~ 92BJSXvzͻ\^BM&ҷcFu}[uA7۲BfKw|;Y@'%PN "QB MӶVsmT@U{pV`)N1ڦhv~bLAw߿ =ǤP~2sfJF CEjw֡X1F) tn b6HNl9{gim-մ|5ߚ=r/ L7L.Q\)QTCM*ڕXďm:zd߃]z!~D1mD֣';A)`z5S[wR fQ"UT=bY- E5??]w'{&`hCKZ*0N^yz^L[ȱXgAh1Jحm%FsS$n*X H)%-V,UJqpDu{kvg9ׂ䲑&"1x{Ac"a1,)kFE8l(&)5>`;@NfqS5"NԻ;'zLm햕%Gdz#7TڦAt/Mcn. . 0l7U؁d]iT>%-|'OrE;Ӱ{4v|Pی&ěܬF̓Va%qaX q;ڻ&#f*L—E"Je)^ N$ W1KAy-Pe(ecxRx?ؽqcXuAo0oϾU/iw_E^^ԐrE%Se$OkPC$c\)D3J+IJdBpgSDNE1f/DVک!$`5*JL vC;l?ۃҜc4&JhD>T4i3܋\c,N$=ȑN w֥v%M!Ja0bCnAY[E՚, :~>uC;ΌYjf)Dqs|3z`=f E+E$~Xgwܢ_i% xE8J>vj3/5"m/H(Ӂ"ִ: Ee/+/K*.dCȌw?9P^HtR˻3eӫg>bw>"]^ Q90Rhoe#e0% 9VVqrTH*!Na?>wwwǭ!megtȲvYr!/Ə,X+ܱH7rK%nJaU5W($1ej &Q!V˚-Ђ-E~3l,|vB˪u$PYDaeCI$PaVQcZDخN>{g9 $^SqZ( *YV:8pfQAܒZ]$U*3͞mIqm l.LypΤuΔsț\C|uV=d08+M8_gYXI 3&֙it_>Ѿo?o?}vrmh=mxSoLʴǿZ)=%{||KM4 -cŸ설'i,LhԳvw>,]'s3{m݀3B='̨>WД+'V[Ya#P GȯT ȺL-G;A`{@ lF&cىnS] tЯ.ވ&Yn­<<廣V~uae^c|髣%?:W o.%Z] Pb֬ y}RD.2w{udV y]ٖWhڐcmb1ݪ茱W (8iYChf3|L.׀P% R}$W6T1;QFh CBif0C oSiuK bxf?Mv՚F%6iV|-q)׿;?772d2跥v>xu|BB~v~y3gs´wA!q!;+m>H- eJ9|wѮv]]rW7]n^8[\^^^^/r|y}w}_?^ᰓa'N9lWG/+IܟK>񬴠h+;ӶlފpFVzPujoTCryKg>Wﻹ=c\XV5>   DV\p>S u)F ߒ|Wц(?- (xHx'X󍄻AWt>Hx6{$w=QMmOvmˮqgv6g٭w?;Ig1jguU` D5X4E0-ږ^T2x'GpoQkcNhrP0Y޺X@!T >kyVWM YlyS1L% kkKAD#dlmTBkk#\5ѢmG~S,T+S>5܁\+O UWKY\&.ss5Y|CS-@yb@I((У92޵57n#뿢yHefS[{v*d+Ely%y.94(ن% <~l|h4FFD|2pXwxT% (ij(1$=9OFbO&4lTsWhƈ!'!w-L>抚}#/AAÛVc>Z!\m7ӃOz1Pf=cCsǑsJ4/U [Uqi-4vHɴ3*G[dr8hI*ih{vRlB!1g'whz!ɲ``S$% *͙q^(!\kTaJ82_GoJG"G<}@]T䙓s)Ap >:46 Rɞ9Fe(w;q~0 5Q |Qj-#2.`LݣƻMbՠ'ܸ |fz!K>xsʓ,|:4sq|/z`q(fE j{5[ )e%R*a.1q;Y_՛3msQ*aQiսSd%;E;-{=kɝb`|\{!!)-&& ص! o{s >1-y}A;:bs=];h ='N?>מAU)LHtM =\EÎT־T10! FFŭx6C|r`4emrmeߜR 'Ч)F ӪDz&9Xw{[mL=xǚ15=lj}yC@/YUO^;fl=b{]@f3O Sw溅xcFZa=gn_do/ܳw&Fޜ]9z8Lt-FWgM4 q7MytpEHtBHϨp#pGNTH#UHZ˜wy78oWu$nH֑4Gu bH1X8&T1y'< ~3|Xo=@@ snA^ LQl]f3ktTT] *)}P%{P1䙋uDzʺtP#~KIe6*gL%5ؗ.JΝȋ&iݳUTPq0x5Z!vjPs!5V6Qa\ Cq(ϟ"J^FcDl&rd4KM0SdG$71xR-)S'*;roO\I§(+ss\!rJ,xo$*}ɯg MbV򑘿G-i3G4Bw fj=(\kKƈ-1}(|c̐3AW5i&Gry`w}"cVM/12$$R-p{KJRn W(ʤFiOփkGЖbZ&^Em *5(i}]~zp֋U} Q:' !AoUpB^G-ݠ^EFA+w~]n}#<\d̓04hr ҿ9/.fb  UP{Y9BX :d;?>yrHeGNb3P$; sD;iΦjjBEe [5l ⵅac$9 $%{Ä">#O}P)Xp}+@k*EoBԄjJRNݯM~&b=T*}X3 Ty.4"/wK\FU/z z}n5ɗuE75Jz:PB[8M3:U3GЮtPށQ:C2WjEgLϟr5zL)B'n,$ MQ_B=Qjjun$jHXQMQePW^4`Y:ΰ:4%1A!S$ݨ) f4c*N]GtJt #}c)Z7vx  wh3g6TMK_>:G|:d^Q泋J;72f:EXv+6a}C{ R9VyO~{A#aAo-/ )tB{*QO3ldLa H3ZJFR%, Kf ;LOA}S3ҤT^Ir,)PZ;4=59#mga/iEj;~8nxmn1-xɷuY/]%#`o(1x"%l5l3~Z:dfM8Q !xZQΠS%hGnP;;mUjߺ=xQJvGZâ]̅ZjgEEݎL h?)Oz[vV~RXl8 OpP=K>`Mu`DܖkSؒfNh̴5~sl~!\᥏];dYu{bHwijGc*Bn5ISfIv7x= 8Q0pїKzME@LjqsfצFkLj#U(O(r쏒F;V5#\3$U65zR?i\!@ht3z}[^~|d{1 1W$ INYhƸkCz4j~wK:X%f.QWXWJ1KK =H^hŃ-IO`@W=o2_MUv/~sYn٫|bQb0+UPFVPH +xryV%1+w(˵+]cHYB A\);܀A1 x4CKB ; %+#pZc*B F(hbc18€U2-)fxye9aa_MݯE +|lsirM3y+?sI .ȬEX̲B9fqZ˅.W/->, L0BJ0%ZKNa%)QŵέNr%(W|C TM(a!)N"d1`*ZR"v\ Zh*]SMe4\2;4i pYPJXsInMP ,o#& F; !UrH(7@ >&(1L\#8OS)UpW,oH*ZMXc]3S~CtӛjlM?I껹uoPAf5)$ҋQ?DlMSQN,#HPpv @Ŷ&+lC$,3 ve"%hu'Cbul; DH1`Ky! =%u):HS^mOՏŖYK1,Jg*DJ- *ܔ,VlXsӲ9yA i99g-CLҽ@=溶(eqޖ[Ɨx,xgqqb?JvX߈ c 7{c@x";:|򣍇<ʲ|cV1?X'R ;XpZIilNgklR^WŨ,U۔'W/\0EB&ɞ GN{9.eb{7+L}ƶlv[*LqMq1$ɟ<ֻGxe@ C#~-tD-At&$"(P44w+Ko5NS u  _"6jM*fP utWܢY+oZ0veOzPVr#ܹM\lo<_3SD?@uØJʴ,LYA+bP2]( ]&G̎CF$!]Z}GpHd5zTX酒O6:s'f5¸Dq1*R+9,33I2K=c+ f)2 j15Uj Jl_ ~c 4r<ٸu|e3ܩ%S[3;Sol^1ZdgQu >BBcDZb{,>Fh5Hx(;J嘸#CK 7Z1\]F]v1Κ:%a4sj)0O/A+ /֯B{P[(22m=Z?ѻ֯@[ SҼs7޵X-Ff-y-^U?]e/'k!$#ж7P'ڎ=:tQ4Φ 0*^A:ju붽Kp 0" cZ5uBLQ.ƥVD}*-&Qs#mhTۜufvr80Qk6-8Mʊ3()xj$d9ѩd2#*U3j(I@ʜf9I3#YRPR\l#T6LpB&ҏ:^ %uYpC{fB7Xw' 怓:/L,0uA@!ϯ_w%ӛme2S1/, FCl%uTnYW՚fg O^@3tj޳*< HZQ|' U8v *kY DA1 1WbuB]TSU>,L9L*G+ThٺrBҜDZ$ܠBA qz4_įPcZ[69J۾*92Ce6".H7!X. s B+k,:=ZL-Ewwpʤr=qj"N͛ta,C;Fwb:R|3<Ց=c(/lṳ}^olB߿ؒ&qӵtgks} )w+t#c0a{)eAК4w 0P S]-w7 Nr+}*ir",ybo "-aoI|EZh9Fx̯"&?wbW?N w_L1^ؓ7^U7|;H˃AnwHuWbBM-=;gS#Fuhy=4U\fk/h h+ݢΈ=jve5Zr9{_?Vi ʈQ=U8FmA iEYP,^YdBrGtP6URj2`DgpK W,\DZS8r^ucT%b! Ā)Kxes8ثK.G3|p3w -~K^x׼c̐SF;9+c-G! ~j}He]K|h]+3=eBHRB?uY-F,kJ'5)#H9)lBdv wmit8t|>T1t6m(lCbBa8р}|gh"ݜ ymϭlGsLKdGBST/tYZiA2fR!)QBgn! J%6 !/,IY`7MF 1 l4eknr)̖)grVf%'I&`sBL$A  A: Qy SҒHuRP^ƹ~Vψ7;R ѽSuQ?ܷVH.C"wc3?|\-+B߱vm<)~||@wxpynJwvk˟(Wb&т\&_VG{33`v33À2lrׇklv=m{QnOm&&[FWVCP[C 23{uRCgKH4HLQ =RS^! 40fdLʴ,LY"ĠY EfЌUiPK 0lj,&q],ԶL>L,&љ2俛.vO/σv{K3G7_DK;"r{lf)5^n-Z^ihWɣ.sp ^#Mnؑ}F51dhW+㫱޲ǚzbMgMS,doނ:9x".M A{ CA>ۛcU4!Tꈠ}*}DP>J fLh/.;SJF{eaNE3l q}f;ixC1FL9%~hi/k1^ȕzJh2Lk?_ RKx|qt&2&HW퇹wX-F}Q͹(W)&* [e\WM%aé#d͒D`fzO3]䛵tvVl;ĀZ*6t}Ҭ2-xJasNd 5\Ԇ2.$%I@0fM`u~g۷# C;8ϒ,f8 lx۟_>lW)NxR]H'CTj@>@/ Uώ[->2G[(C)&bQ1X|H%\|uCd[cCֆ#/ _2@D'U.ol\pV~pd dgnZ_̠5|̅Quu!8U:M\LɼP$ΫyOӕױ!:z̞ʨ;*첀ϒ 㥙wr`4B]gdPBjb^M(El\fzFֆS) Ӏ (N5jޢ缘 ڼfpq< @e(qfvPg%Bh 0~JB^,~ߩ$Y8?"< btF5QXu cE綍d[l4%=" B# hWjӯc!R-G5Ծ* .kue UY~8)=Bu"G[s[ɱ~by^qx=\DZ~,{$r4Җ[Q{zZ?ѻ֯@[J+J\lok-]b\lQsS[op*-!%edTvudZo Xxb/W@"^ldxj.ф/6[pL9Mϖ޳&!Q9&Q([ThBɜny,.i:ty(#x1yy1EaZOBp͠e`"RsJXۚvR9'G+L)U8U;>4:80 ǰP%Z, JE"h+ӛ41V ۘPR$3*%U#Дm Л .Ir@vWBUX%D 9Q 1Dg&"h ӭXS;jpE^[s0L"L!*H&9Fw4jG;)A9Z7Ri"m"meb_g5)%S^3MӌdBd: v!rb|LW4DǒYoyz-BM/gɩO-O" .zY;yqEZha^1NRph#«e~F:kmQEqM]g-UG>_j:m@-mNvAI0A8hed6!*b3>L098+s ;=9V<xÙߡl)߈^:|ƊC-FB*&1LJYiBL"r )KhFi DVl'- Z~&-h.vrswy'7adm2krw~y^}nDžWg;oU) |QΧWݗb~{dYqg/Ԉ0"čntul5QLF/kܐR/Ž֎$ZjɲdrtL )8K$aSW28}^fh홃DY9; *R X/_;^yǏ.:M8Fli4Ս.@G';J'G9 ^&]: /xB:|тiZ9-3yqwFrK_40v&"wp x % RUS y<\P.G]IKpFɏBbfI3hX[4Fڛ> NjA|}GV\ jޡ#~#/}󃮦xlџ ZG W-=Q[ɵ0Ls^ǖs@SVRwݠ`;<]RP.ݢ$:\?F-ٌAUY)Qyoք⚞;d7=JlvP_<_]O[4N&ףEbX ;Χ^6=e`4kpZimkp;OM_xQG'? *T(Οpⶌ!V9TUZRXl N|-.Tbfjbu#.]ͥ%Z.x4 ֖utRaޮX'umߞC P B Hp::AEj#nP)?P ~X'3GLyQ1$Tnt.ei3$m&qQo%̵ H#&Ԋ=(zWŤn018].Zɚ5Ww'hwr^?NW^Pn)`YEqk%ެ)@3Eߞ h;Qa/^ J KD7&ZH#v5F/xOJMj]m\lis ɝ)Ids$%boͨoOuAK n*nD!YW7B {l 넷}b|Q(D\ G 4l]kS9W x8 ?2jcfȝxD% a&B w%юgl,&@+ $5yRs1h-0|Ak<97ysB+@8H2Lkufy@H ^>2҇& d,¨-oGH+d{vAMX1-i M$ݐb"AL5c9hC`PkfA)Ӿ|7uAFAFް|5/l66% 2k95S<+7Q#"PɎ2Bq܃yd'D34AgKxxzu|y-n zZ]~MTkwF0z*e}]irY_*H0~ZɶF KճjlcVe\h3dp, J`n_r}F-#*z!z?Xm]hTn(=HBîžo'vn:@MBsD|)eKw \,oJsܹ ׯ:҅;mF>MBw`㘙"uO6 6igy߬~ǬǴKGsHޛQ\t>|־Va#O :tJlJZg-. Ys=I y`[lQf9 Ge~jMlcVz) 5kerܲRkh9Vg{& .l>|Ҽ|1|ef:UOv1P axi KB;SlYpsJX@< ! ?b?|᏿;<=$O/3`Aw1g"4Z$H m)͘G_(_Iq({V\>;+9Z+6!I`sQG$c:SLXPKŇ;t1;}i3x܅MHWidfpt3Ȼ}:D42} u?C^1gsC 9C||u)U+CAڼp}*^r//i1JP-%+ty99;W_TnU.i?8o? ˢ|qpvk#(fM*2ixv>% yl2j^k 3OtTO]TEJ$3S\F@"&SX!: -hX#{Qv)ɇO '-D'n~=Hz^$?9'Ԛt~6ޓ?[aVzK -2 !ktZZ=VOiӒ &ޚhJ?DD)Zr4mAYL=+NaHr}|E>tceŖ&֓ߩEW.0;.Gѧri}ھ^`**. N1I $Lv.ɜ#)X߉z-?_T rOm(6)AN$B+h}qsW8qɩY!'mӃmuA`\5Xմ"F|`6:MI6'm>i | *I.xqiGZ`tZ$Y,SR7|} ,e#0NR"v:"\n$+?ӓx}y o-p? NkEn̂|8[mc譟ʺHirK3=l??}{DFj6e5YnL|>6Nӹpa~d" <ֺ*@Qj1MƏn?zA9X2REϒHW׉,Rzk}^|R],N;Ymg\=j@dJNgṇqHm+/x6fCKB gu$3ت1BWX[X-_+I*rl%=i5)ӯTvk-_mK5KуոAh$InvJsJX& =kE)-hß/492Ě*IQM.ۚPNΒfF 1Pb]tL9;a!N*_ R $ ZXm\ڑS% tR=0$EεJզlJm* lhWj !yث4O,Fw AYvx,5o@ g(eقNy9賄'/7:1 < B{ ZhB%aIsd>hd=&=i5oWү/Β;g S<:LPYL:-[8 ]YoI+^%}јaFV{(R^,b,Ju eEdDddFD,b9mHMv\A`;u*Ь`Ze*"Ŷ. #IUJ9HY`#Lf^:(;r9 d)$[p|0?Z?ƃqP}pa’' lip;ޗm)hchUh?B6zZS.NFx+l>OWt}h~u]L=b~_gyl G#1@= .At p< ^y-oc+Y")S -D 8;6M{0*>9`sR(qF" &[C AiGRGp+ KA&1h;fPsi.[',GZe&O>;nh6$>ׂ.̚:2D۲eb\Ɂwo˖#w_? @pmO)\gxME.%ևDS|4%r.=mNi% ñLO~ 8.Wvɫ¹)rt>< (8|-p@?of'P`MZZ2 ˝XX^P!TVNJ9,epr=fe|ح}d7 !ѧqz^^ї|1TZ]i8Z߅)Gd()s`@EpHIp9E\:mp +:[g8" h?5J\_} $fH* 3 ^vbgzбn-H"G6%zD8APjϩDTǃJSka^NN"ipa5x]hƳk4US&U \q0Ʊѣ h8֑hkbљs &*r =,ϛgN țY$ɬX\Jԁb¼DpՌKր<,a hBRNq:BVT U`85ZmUB0 \PC>eaB`9I}YE|s='sקZ#.x{5|& \|Km𷏿,¥-KӇ]woflmFenFs^Il7"'[(th,5@\&Uhn.,/lx=]NO&.EVqc9YjjF 5#)N4&Ig<68؉9hB]|+MÑ:s_ rzVΚ I/`A2^ ,Xy~MNjf~e@$wcKz37SILy.Q aR6+;0m< rQ׾@)tbuD+3vN7>m,NSX84mB:q;:zC3x"]3;>'a:Lx5+ِwv%ɹmJVi.(]tݡcL0~*Hxa*o~: 17+=_Ə[Nd9G,OBHA KX*a DȼRA WVuUp/9./rc4So;%sQfj)aILN2e*P2nOVZxNqS*86RmMDUisLX@ *Վ{i`E22%9D,rӨ s;JqKWFM,Ơ5pnY %F^у]=CI0%.$)*1 c4e={y`4Bcn4^`D * RM]L#&%3,{JDx6x _Jj%O-cLEtߍOw7QWbWuwf~suvL/Fb8R.V +9yv+ u:֦6 Tg-sQjCkKQ2NG5"6ρ)6j*sT*im*EޠN mvxByVFX) !1W N 0 @Jga&Tσo[WB!j=^ -LZT*<C ،lJqN TO`8qVh֧ZcNW?RMܬ:3$_\K֫|:iĽn}lzgݹO39¿d1rbCtcoA쾽sы?M߱)3~w@ďY?S-.e}M!/q1r{NtYhLI6KYRTԚ+{c=X-W%7nա\WUapԵdajGb2TZfbϯE"Wgtg+l1f4J]B%ςP-8RŚD֡[]B < k<&ZnY[BjO&3ӬB7F[N!tl:D jk͢קLj}SƙOQKRQLhwf>쒅jaYi 4^dAA1 ow$$u*˪jR DTz.xjTA$>wh:O)a26d6 BT{YrAtELj3mB3˶6se$.Mfݐ;Hpn'͂/>ȭ--ªGD'MfwjG߄:*xg赢(FI4aՁ'f1̬ɹ f'fQS.tɵ`}&T[xCUKJ`v+;v%w~Xf(X d-q&|>%=Ƀx& |O]>W!LA>jG ߾sϠZISpN@x4Dj$3u,K ' (DT9}/1Lj22-[䅒 $#{snW>Lדd 'R^Jxdzɾ{ᾤ9Sҿ)p/S3{u aTˈ9O5&O7n5%{܋vAh,+:nzƜKZ5V2~l|[&*A;[H[3Gn p҆dRG I 9@K#^Bjܷ+8ء`'s0Oe 3h}{ ~zZkY+лfXbL ω&\&F(SZiN4Kkt LRFKf򒉽zJᩡhӀNJ rRQPR0 2^GZkM3Q69R99̤j[Z$Ip)ɤ ^ʽ>UbDgDY$ J Md lP^ `!ZK˹$8LkҸވ_iAFk-*S=;@zt3lo!M*o#Wg_3Yqam(kh-Fed1[?98V7޵8r+׿"엛䦵d^ @ŀlwehƒfm'OQ֣%S=YޙUYNºB4c0KrsY7OoYEӒ4[VѤSY7btŒQE5rvxm=v$&EZT˚|Cl@;A MddxUO]a :F={uvuw t`&tМqCwѹ#R e5SfVm:hi.ۯ^'3p ?`R\ܽETQ0̀{:c ME 1&i2A+h3Rf펭DHmJKkA[t~Ʀr?xl+,FS-37'LJ_y@ g!F*dg\\CyN~S`zSr5t>`)RpiHf 6/ׅPtFapph~0E_`L<9R!JK#JXO\hj'r<+x7{h@eFlRi&uyeI(iIכCp0wz*h2aQwS5B"&Ip4Xy^H$EPj(Jqww $nY9(S3 {r՟ z* ɤo&Up=z0 v^=Ps8K)>[cokGߡLo()izu "P@2VmSD)+!H@ J9S`1.kja2}ګݺ'A'yҟ[k{עsZFm^qr497fBZs #R5D.u1=07 FYjJA5A(hNa*>^%‚zw̌VJҀ]CB),ARQ!8 'U04/85e"pq#`zI>Mk1>i%.)k2ˀr!C8=T1 (-x%C!nWXP,A* Tr&cz(DnfTseՓL : 8(Ycmնu~7Ƽ\w?T -g gJ(u?(S4t9#γu{rQI~'nϓXBdX,P.ol"<ߞ-Ïſ:6P* ^U1UKDv 0/lѯ87_`$ pg-o寳b|ٍ_,3#> x_Ih ye gVHB8;BaIfg~x 󐯞|YWj[7OF3|(cJϳ"7OPZǵ[)cooa_[YW_fE~qӪ(O2}D'&:7щ);7Wzudey]aW݆bF"zX1 | 28>4P0Y:\lN=mW%mkAcE h0>A@9kB:]j;m52E5h)!P'fv1:}f뮷y}.npDؔknt'M4R"ՂL׌{nA.m \ 1HMEln JJHL>} 囥Ǐ ak8 ϥJKB0JB1w-RTr!Ҹ8,i%TJ줇8ROCJN47'X6ObttF ポkhN=B.%1r{zg\CeWL̠voE8x i=WC[Ri!N4 {% .+!`jjKgmS;u+Q=ZLGܛ/imYh{Z9Ӄb 3 7o 65L9 ZDB*jI ڠ!lzeik4kQS"H7T7~UtbkhT1ŋ 2.[Ǎ<~o:-1kTĘGK\?Rˋ5SJжM&L3Owjj^[[7NTŀPtVtĨJ!;qB-ݮ,]0эn'eF OYGL:zVѺXKb_ɨ\ MZ3t1"+T;jedY({Gېow&Ĝx)M> `ŢY7oAo6_fm3d"K[i!dbk t9a~=d E/ .NW-,f1ů|ڤ뜈|Q^NpA{ "!@-uYH)LI.IVƮ Ru׽MDTl8u4.`JRĹւ yVf`j.Cu^cɉQ poUƭF.vvz+?6j 9 lpѳ2s+u:ر.w2ea,n "<1ͅf@TJIlřRt@gmS\\tҘ8le44 $|݂*9Z+Ft9 哞(5#<qu=BS(>D^ '_Kc|_tȲeM}M_^/8d*%1Zp?*mN<I& <=؞Mz80d#FW!D5ȫO5%uysr»թ99gư3f vNTC[6(א뇏 t8oeKaI*ؙwl`mrđtlr1" eq5Rh)q̞Fʆ#MIDi?53AtkLx3x^Yc[ 8W,<3yoYE{QY@FkNP9JP򜤖KщVf̵ĉʨJ˽-.7DRUkB^1903|-'׮39eh# z@1~.=b &0tj݈NRu8]1*9\K[$Uݵ M(ż5Q&I=Fp=/!C5W$c{sc# \v+8B_/n#&|͵jž E 0"/i5R칤!=~)NYd(T2I@5x1;'f "<CT&Z*'GRBk Q8[Kf m`LAg # 浑I=S6I"Ku~10„hq'8{Gg6-n&$\9b Q7ޥ>uǬ~h.RC3(SBJ@Xj ~ΙAI_i3"lD%/QckRTh aB@-1 ЎH4!]C .jV}9A9[^-+I6eqF\*eڼִO튡0[L$7(v3r@NϨ<((/vvAaq#E1ocq.ogyx*U'˦lM̦lM9}CgL ^ rZU]]]TP ia r*h%OIUr>-:9~>OwGA+A2T~$w bNHLsyfu`+XrT=(y┪Apн|(tM+Dԯ V\hE!-䀄ʽy޽"͚R_d l_@s5,/K+AS1V rdlz{[R$ SuZ| B ENbqjԨdBpwH#\ &mC˭^xB K>h&UI /h(Kw4͎Ϡ$#>C/J(UWiYU%♊Ģ{yY5lXwgW/fzq=./LI'q/yy)u/F˧ܡ  6$,J 6 .h3Oþ//ΑzK SjVlmnJaq"\p>xd 0r T-!FTY驶V[e(KNT` i0c\Gb.J0PII#\8 s<SqYi\'8 JKC{6A`s0=c(xO|)oS@bYb1O,3w>V_$QH<[QYuS_}5'#?g~t|ϗ#B犯wlxvoo(.{N1ךԤ<#XkDdѢ>| { Φ,-k-v8/S HNyBRo)sKUDD9*[RJZjiƅ$ܰm[&i(rVp2 )MRyI&7{! 7 č3$(kU1.GEK gwO@QJ;=\r& "4-wFBZɚN@kwܔ $ȈD$IZ7*V#4Z_xZLaJjAWdYʋXS\(VBa8.T'D9LjaځüYjx$F u<?q@q • V+rgSX,E.t)O.ℯ8mp*֝Gp'j #Em Dm|Q[knX\+Ip,j5w F)F $5ƋLec1H ^@32[U琇, ol08)a|h0*2 zlc,q񭷶+rB\yC{%]ݿrPSPqx*$k:]r$\n9dL ClCfO҂+{t{Mz%ih%ԎJ 1DBQWfWՃ[i@C sAv((5Q5j,5j[֣Zn[-͐f~+f {4Gc9 Bzoߏs!&:}N\6b^+Jc4B0O2RkNVE$, YJ!rno4ˇO.cAþdX:/lg;{v9/GZ[2ؐN]ri &Ahgb` fS8gy1 SsaΦ0}ӂhRXZp*.us΁^"~T|X&`}C*" ]>X d˹sVl7$@xf 76|.8Q_S׿r Sn!#GH`"0nWy0: fDa .K 8=X?fm =q(VvH hP/Ý_>.1kt> OBY (@-3j@ѭd?gjUA'?_SpұYo<|T"܂-=G= x1~TomH#&M7+dK He̙sK)3Xvl_2J!qY;L6(gĊZl, a3pW{Ljhtz޳4 qTNqapi UY0oqW,(*Ȟ& 悔[dBu\2Em$.x^6OAK{_6%{hpd/ `A`_!팙uwJ˫$ƒEL-C=SG& }I=GI_#U#NTtZuSzL Gcz)kFi㊈XF"- D8܄ %PjTrDPg4=3c:G#!7f!;_%sPAlbK6[Sq J"osTIc"g]P0JIiiF&fGL4 qG'SMPYF9GV5>e k7ceNRRhcNh[]_‘&H)T#ڠ,f3eQi SyZJΰV"Uj,,<[2G"4G 7b n `K%P0vnl'[+s Zk!h9+ߏ+FV4*"-Αd2udwJ5c"Md 5ѽ+i >Kb喇{?P[ 50m={wmݿ Mb 58(Luu -F+`i0Qbi`}Ԃjvsf Qt 3i%Cg .nAP=is%ɠKC+ڶD9|DP_̢%咱VHw=TX^9]pkc"qYPe6vG[֛w^g=ZwJJ^ZN.:l>k6ۚzo?/8P nXarPփpШ}|:Ae5C0Vbd!DlWm{7ZGw;jԩyQޭ y&eSZ >nxDA>w;ޒNiU- cwkB^fٔVvR3(y[y[iq$Mqqj/5|-oa$`ةՂԴu5~`S ' RK' m,OivYL !ңOvV~TbuZ$8G\=>|v*;{~:%߶d hN|@sx}Ӈ{o4弿6e_/ԓ 1:i@ZfW Cwuee6eZf?1Bԏss3r~g"»P.tou8՝l?۪5νcbE$G գ.c(FK,SL[/.Q66{khgGS˶juL5lQxRRc\y'ou)g;_C_P]5w;rH8i)´>ZscV 5nl?EDTi?3\{rPu>w'"/c`(A1q+ҍ.6v1GRºtotT!b喖5v`*;-n3L'?+MVfKXZxNLUʪjKBh0{Z#W[JuxQҟ"LHDE_`${Kѣ9( d\?x- (Z~Mވ)ͫ?&(*Ҋܧ7eH4E\P:W_ 7!q8X4;@WP>h芮l-B[~*kEaï}xG]3ܕx[ڀPF+Ub ~"j'A-\ 2os*jַٻ6n%WXz9UpQaqyJ0 EiIʉT6e$`fHqŖ З=@]7||3ykfHUKgDDlobALN"Xp.;Q&tQ^jJ sͱ(s^ @47N`%a}!\DyAmF JYKjm0h< _|'5 M]cZiڪT%b>FjoOwSJaYKjǍ!Z*qLƌ-Pl]bClyBQيjJ'K'4B*}H/sEZF"- 4^[1| ݋O2ym.wpiM5J]s^(N|`VYl 9nѯa>Jo_7ېL=B1JfJK,~K|O>Tتkj[楜`}x4z=ݗ(T{(*"f83RyE&Q9lYܿzM̼j쫄 !N>Ι>yNRaGwv). MVH-δ`#+Uibc{K)y|""lB\Bz<`Vr WMppQhcJ&hH5ŌҘO? Se v9j?9a}C̀3(L!o `)[IF@PP ƴuRId (Q\tag܂ }sOFyp`GO@ ww_# SQ.31oW*_d*\&O}#fV!m dTpѴlMp;3/8 =8|1tapfVR H'H;K8sjP!sl`{0n d l$6`{:D_,+g>{3DUkws‚z*B̷+ q 5Y+ANjw p3V˚?Y5c3^G֦ڶX۩ViץR͌_O=Մ[ &1#DeTk3-[.(Z*%93%gk78 ^-1Jq ?YE:wopp%\c[zА?؞pt<_ݳ8;@=JPҝJ~' ׼P+ .]yNaN"gX|K)[GtDrVLYlYi;)T3"l8B y1)25sBLp {7ҁ|(˩ERjtglwe'KO\ ZhUhTauJdFy kŻ\R[t }\`;G)kN`U3Oހ^B5݅N/ 6q[^i KA'pzc^`A%?KDUvj,dbSZ,u&UOwSx*`o"u!!߸F-Ab#:c4nOL%O4Vu!!߸&[ڍ)IAb#:c4n'pY!Ș1:>g`BS}>j懡ft q+pJ<\&? ]3BNT==a֫6QcdtV߀J8RW κ/1l9鎹%g4=h#Atwj  &CTݼiRs爴/)srxpkc%N"f=EBLAQݣAr WK"Wj]$j*\2< M1U)Fȼ: xa1V1d3MCM:nOƢ%nenT\o!  G $ u*I:y)wtMbՊHgU͠r:|ERK\u>xϧ$t\).87х|" SPH*mVw='Af>>Q2 TqiJ?b*9GWv<x ,!D%bRXBIX~ ,SmpNNbLy@k3#2])"s2Ʉ̘=KLeEv gxm QzmH+0ȅc: Eq娇!ҥ-1IA(ś1.=aZ({oj(F=m94}d"'ٷmx3f.5y(+ __I[F01#v8cN +: X⎡3[~"G1j+D G!U{5B6a c!ZY9z6BoШ2sz>[1| Ϣ ~^] rwA"Ə|`%uaq_ k"(C~EA PfLpu3 nԁnF+Oj؞m׿ X1pLR WSW֝^2W>s~QLuL+/sLe;˦yDz%ˤ@r@@PTi]~uop "f\7K_7Z #D$#5 ՓC7ۖo_:b WoՈ#^CU̵͡>D4ƬOvbFHU@y"|[͹ *p\xC4Q߼/`O&xվYX*ɚ|;gpBcQy3 Kk &+J :pz`9L_jcoX#5A>StJ›.S.=21/x&IX^uK0aҩv~8 ԣ*=OE_?<-fn<9CK{Tei굞#Js/B\]꯭Zm.V).VӸXvjnfQ $$gƭrkbpnm> 85QO K6d14#р[q+4/[UFUUhnpeY 8$8nAv@chlW%,mCaQ%^':}_,J3PvݔLU>SVYomԛXs%dmsA> $xwd6gRFPNtA_wD2X/D*%X*\N +üM] ]9sLX:5׹F pP\kC5[7T`ID oJwڵɅS9t*S](2_BW٬Pnq"]]6?q$J֟ߒן_U.> <~0ADdf,u!5WbY{M/~Xn x6BKۣaN5Hѐj2ctt8J)^n$#Hړkq'P9U`Qg'KpeweW"KQNUG)LtO4jLqK%|uIJ:GXQ9͝>.ؖAV mƜB!(>=VFI1I?‚J8 ̤B:bn.KjJV̢S3_V\OΠnN+nx놾pk*(6JFBτ0Xj ֎$"\0@sDT 'E ᥒ̪wQ!|_L$=cTsqo9A,;)DZhFPF;WV )H:Y4(‚M.4o.j DuN/V $wqi,ڰj 4~,xۑ.hBK zb"ZC dHm`<R%>VtDp6B9%JNԣ`ItE@. 4\q}NPuB7^Mbt*Co+D ft(D.DMlS;5p)PD ǶE@/:FBJ:9?blDڞ_H_- ^?+e>^m5MX$n`lR4ϼ%4`Qz`JۯQ+.hPSJ J{%t7P{I Mx.'6؅U6 IFq`˨@ЎyX(.%?T]NF6+'q;0}^6t~zn ۧjJIE,L6vq iqD Ƨ tCtf -F.qiuLhr53bn4$F ѕ]'kiK~o8[(ov']s I۸\VT !; exr C(?,Q_%_e4Rtz=t$7ESsGSq˙5Œ6eKބrE@DŽ!OǛPl^X.3WM(.>0uᡯ أ cPb=t ǿW.d4a GO}u[v< RĸC1y7HN,Ĝ,Vܗ &V U@94td ND*gQ4Z nR¤2/]fMPgI ޏ6\.W'Ь:Y'E^s@j7i5M]zܯ~z>5lR꣊Qej؎u_oX&YT eM ;c@iF !e!wTˠ Һ).z2uHSTO)~'sJD?s ƝD/ N`RKf7.dZ/PJHƕ,,m[T% Q31:Fvcrqd>^t<՜&gq8~M#٣>)^|gkgI.3hvwrf$aͻ';Q2?J+5' ~>[xu&㍽=|U9kD}rWO$v> ◄PGǽr~0KviWJ73LfoSJbG$5[Wp`S[ykfQwȫ*Bz3xkvM z47tCͯסY0r~i8>|eJ}bRz|=rГp)A'T&PbP":F$jJGn5ڭ y"zL~[ Q9'7 nZ<WW{>К C^  h a /Ζ`LI3 \'S 4}j&dKBaÌGx|HsB<'S#8YX9:qZcmSJ}jr !yŭ!`UIA}ZOR``ߛzc|Ef78zo*P`NNGu 6 |uXj1L Dg^AoM<1,ĘLaťA, + S,`TD™U^[ɬ&6G. 3";=a$BYģNX)JgL5\rIਈ|}]hitDHkhPi*'Z9 a R,)4p gpn'76+D[aJUa5t[j| ۋ ʋl mN^F(V},2U (jmu- EEZF1mV2h;Q|ZW',lSIhѻXU! qʘ>kԥ?gQg# 9Fg/Ή`r:%ˢCyV@!x5Xtof F2^y cH,rCg|4CQp!L3[/އU'.Y}7yr^?ϒ'ygy<9?a$i=ΝULM_vh.d Q!^ɤݏCZ;r>Nrr_PFg3m408g @1?[]W7RͽLM̅sE|I_DKuC$1v&0DYhqг>'RPA *!5c)Wz m!e i[})9iř4y0`нvdsJ= v1KPS\ n'B%cCF>8@?ԔG_*NJ.[ѿJLxK.;")Qp}Ǚؠdf?ާ=}ާ=}V@uw;k\@  ?y&P 3,(H xLj3!UƂH䏷ߦ.ZFo/V85$q{ 7?^6| 9v8y02W;U^}Y=X,,|w~q'Pbz!,sggp2Ț@NïW|Hg[0e͐2}3q5ެyܯFa-_1Pi*DKuAfիe>jbsťҗd5(GѻxGOIgbAeT1iӤˬFgOWǑv=d #?`Zʮ5^=IfOL'l[ܔ7'-nNa[HCu0`4?i8@~8Et0` Ol48``Xo`aYL" O7](Yqt{zfgN[30fNZۇ(bo>e) Ls$2t'&F;kNE ǔ!n AkuXe=W~s[rOܫlaBO[]q/[[@gbhneԗߏΦ/7>Ĥ!zĔ㷐─գ3S1 TmuX-|B+DlV{kr.0h@GYv |oMVG T\Q[4ĔKq2pM6F0Fd):"8L uP y5T]§:-1+\VG|OI$Ձyojk@ 66P D#NlUKN)*ϸ$v{+6_RZZ4h DJT:2J6BAp,(z@*sI9ojjSmͲ~@{"Y4r_Oz ENH 2c{ԄKn-qucˢ`U.I۱NΕ㻿;$kȡճ*}rr @lHԯaӭAÞC R$c9oϏh~\{e~~Rzuqr ' y"Z#S F:i7UPbP":F$ڭ~nvCB^Vɔa)j7%[-%S{.tqSCFnS [EJL>"݀)hZ JD]| Fn)ڭ y"Z%SǞU?jiƝv/ц7J瘴JhHѴvC`mӑ:O֌FeUP i]O}.r,.ϴ0wP"|3$wY],n4}M >}0_P~¸NQ6't YY_֡U4I/e:^oܿՇ_$ﲂeS",p3KփcD,-L%NŤUTOӮyҝdU=ਙEW7WD=1n?kx*QU<~ G.+8S/Їj"5MsMRr{[1.E `ϖ;ͳeC*N$}*;9gH!fcr둌*BGK:Ut+)7L/XD,1[RgԽ*/-j-2ϩ;R%' Z-o+5_ƺ蓖2~l3YGٌنsP"F쓖5*Yul?/H.2#*o[9U%~t:v[N-Ԙw Mt!~Ddm=F' x6qGTXtĜ*L̷+`ʫsi3B33Bl!``2k@ѕ VERA%̯M5j{QI쩳VԮ ow5֯&TQ} ( (,-h{̰Wbښ<;p($. :.' $t=,T:TI ՅiI4;;WweIz>Ѓ fgѽʣE-iHƠFYbVeU$6,#8ٴl^=#Mޘl}1x!kg>-X<Ҳ EKfXpH a9dAXPJbl ق[\ |Ec4h ZZZb\{ I=PP (hL/BVHL``*L" -gW^\PdFQKBf;F`ƱErgE"?yDOb2az͒  B4 }p6'C,D$1/iТhK5!33@?.6u/6YW7̖׿WWѳ eGiuԾ`?75VG )Cˇv)m˨ԥ砾gE Y'9WXX(OJ2xL&38^08,. f:{4bej"0 qiKw=IhP|"hzoΣxt=q:<\$R*Ɩ?gE?zPZf"N@W6Osc^3=6,9ǒo&MDʩQu *S(C (`un5+re<7`WV , F0#,gZ#0AXOR`+ # V -OE\@gUH#x^&wyQp@0՘j6Vj S㶟H )A$suaV* I.Nη'IZ`IE.Oǜ %j2P0^x8Cdcʵle`㺕&qNk?,i/AsUD;•Q6-2KL/GfRSh3%2(V[e( $KN%F\NY=-ⶍ"> Zޑkө=H6r_3H#25- ʶZiYd6vޞ$At20GDR㦮4ӌ-,fzp&|gC5(>mG"+ \F=/cqlԂ;ihC{6A`3ҲYl&PMs[4)z0AtD&YJ T la?H/#]8Zc'm)}Š1biz!"Ioï5-R>!CqqG јvF) 6Z(ObVM#&:\(K:A\0;9Lu}u[MFKau8JBfc0Qj-jZjAV@r% KE(cF'T+#\9*4c`5_nI9|*oCks=ذoo|~u1e-I\Dd6T|ݘTn$N;ji3L\k@TւELi{^"D~yB:sϘȢCL0\^sv0%.9Hη7ޘYmtӇ4A=!maG1hH4L0Cj>ЭD\5DlNF>{(r/g}REZfwt/b/{饲VWj2A`Jz]Ǻ33иBH;wmQV\; +h蝅a:$>2w@AS!k̃:Q]Kx3\v ɗs;ńwPOo`sp~-]CjաpoI)-@(:$c)&w?i]ga4 KC2Bڪ^!iclw:0ȍb泿YCVRO /ŋRf f0hV[̝"gIb,ޔʊ7qVjS}v%Vj|%K8P~;Y?vXx0C)<ܥQ2ElAӊ,{fDUxzݬq0I)ZRRrhy~S@~w x}<|aYѫ1x1y[ wCl:=^z_ls%&Xe2~> DŽ nnvtWb! v\X#R)RtGY >Iy*3y~;vΏg I<8dq9],oܗtl㙰@ӠJl8>p.գ2O~?v`+,(9uhU%k7N!$W4ZwKȯ)IS%UK!OCİ^xWG_"a׉hˈj-tϚ'@$ ]m~UUPSdPZO* ՖuOFkC { GLo>ba/?n(_|Eoבi*5&brO`^N>ŷyV~aOYj gOJ2|ĥ5Wϥ1k6Dƭ+]]eZ6S4e|Dj%oWMۙG}ՃX$tULWe9\Қy29DSΞK?%pBs&=C xl=OO"qNw2 eXlL5=ـۤjXʝ-muWja[VtAG1h9^D>ҜUw^&WY_IO3!7IVYV՚:eJ519RIWSZ; \vjpKQ\NJp9(!$POP jQ.NBv\QfmbצEBnکBE3,Bt/i8yJP H7u$3un3JE)²[J7r, 0AF3`~PcfB Di% ȕ);k=)a$~"fh3-}xE }f}5SST?LBxA^kBQ#/ #2ṾfZ&W(3tKL(Rg()3$$KQA0Z 4TG|_ x&:`t@Z[R}eqM:k2 sZU,<{2zR[u){pAw"λ}%+_5tG=4UBH %Њk[SDt_}?fnt;^}o.v5ޙ䋙Ofby$r%d\->SOyB1ABbnuL֭9 3oX; ",(*}5cT ˌ DjI3iJӮ=e8r0Y Ygc%s*+9 ),@ XQosZ 8pSk kpdWR(G2A*uB i9vbT#A C'FKa]`+`k$T;ĕ*ʅ~9dk-&97en裡w<)(.?L9E3v>K $!}1hHIVHLexH+$p)J6Xzaݒ+ jqckӰ?r-$tMAr\+zVlX~G˻şֿdD/~)5fMlw+V~ΘtM0 ~(+mwc {5;)P9ap V0K{PYO/g g.`M/t߀3*fW-76xq0(cH?.ޔU,bDVa0jNV@29ҸnKBӄ*O!J @Հ<:1YpADo ؁R)ڊo݌jgz.f{ o,:D6\M9<VlKzm‹ҮȌ>A5}qXZb@H±دByfZo,/ lDa# 8r ï`?/(v^8&3$aX4n&zGuQdipu6z(SH-o6heT:LPaXp2&G=T,4O:Q&۬wҐo(nYxͳ2:h\`?I_W0A lQ VJOMFpF0rsgGTCac6],.ïtɘ.tneK8受8651W`W/m_X^\ W?_ǥRMշhh~tx.Vi]Xv_Wnfq=+6]ӊ=X'<Oz{'CwWm'OTuCDkg DVA[ .f J]!)-,ubFQ-a˯YIM"9cŕ 걣#;IJKWaއ:qq;; L(;\7 Ne-ni*Q/#_| R#{΢V>0[)9S.bӚ6-݊'+΢Ev 5P6RAG,ktH+J cՔ7L#ȯÃsNe窯z,_s|Wh2s:J'FI$v d6pO;ݰRQ:JW3TtGEU Ƒ"N2SƸV2?Kn+"L?wT^,$ Q^%Q-w+4V<YvB8Wy..V'R%"V*l1T_Z58b %展h@XΓǝacƣA8*M5 ' ]P:W"Z7󿑳 Y!CKi;Q*g&p;_-"_}1-BJ쩟(RPjДk-{,_? %H6p˶z556nS`ʕr>]2,Ŀq)`{H~/,y}>?͜Qw4mz; G_|Nkoޜ0r~dIm(^Logo/oNS:94siog_/ss绻+|>F[RR@.=+ RL(:'y7ك>2(3L)mJEkҸY>(E:if#)*OΡQ$1_!_jAu^ܢ~;!:̶2 t@y*q,б0j@`iθTYk&_VXcdELu6ͧVL/:-RMk-ɟ} aUwf] 2-hr>צ;^FUp=T$QuѨ=uM7;=pW9pxI J=6M^dN^L8O盖]jɩLVäE aAS(5;I+*˽6GNP,n<:J]bh~D*.pv ` B6Z)g,3`]qHaP 88"y!=邌&0eP^LT`0 r$&Get|%%/RQ riVeOjJhwiwh{4 zP |eTthŶ1 1U唀F{e!%jBS"bmUTE媈~שLԄk<nHMl"be{: `EWD"`B( S}$8Im&) 2K1haA0. # pذW8ȵ{9B4+tMZW ~t3nTQ }[,DZ.A&d mbQr9'B;ĐnVZ:(T^OG>ɣHRDT]} rKGeٍZcRK3 I-ڹ\B5ga=.OVà$K*Nw(U^^b|?~FE^3*U FAc'3?He(J?/jij[u+ԝ6 lqX`7 }h`T]CFu)GB :KDQtKpqyFv2(\ {T[\'GB+QjxJkLZDk-)6}ώdQB!6QdڤRI:y7.& F*m/2!fT}Z Ӻv&k O<`ZWZ6( $|O-Ov?GC. #f}5vӋy;mݱxA*OJFNꐻSNF7J}-ޥp'p@<'U0rh{o6`%UfHlw, !h!M?4~[qk fk g/{ù}N+Z&eˣ{_[9.1Le0 GB,۫e}3R^*wB< [w+k+%k-ū^ ZY:@"0AHXLR D^A9wM GG}PR S=38HG"U'_ZRXK%LeꖟQJB"ˢv/}Hk8NWIZK/sA7ʞy3by=3/E !W0}qûXrM~գK!zx&]2hsF 83HLiV-4hdG Ssw਺Cpcgޜft4ݬŻ pyϓϋl9#l#=lN^N`B-mF努.@Ov*ѓ%K_EPr5p"oNCh>ߠ @>nV*hT0bvھq*&ۣoeji5Uť}PZ''R+I)OM [v2^&6.5lٻq$W|=;9@i쇻~@Oc{v~_e[)h 8XzT"1$ ~5oꦂ!}K[.d]ʁ'8 F81מxpgLmCX@ㆴNAq itP0.}H6UpEzZ" DUҳG5QAHGUjBV8TU4q$ܩ@ mcwŶꦘ%ʎ-MZZK) i-0P4|>=}QҾ&fJjքܣT#bw)M`#9ǁ:]o{\9e̥Z|g>+|+n>dG.Lu[Shp [Hzm A`웘mTPLbKiNᥴr?9V[S6!L[;Rmq%P85uZuw Ph| >ʶ;א԰ q V+kDPwq^CfOW>HK9k ">A%αY NW~"LPCNy_ g>E(R^wkzP~ڒT z`LnHa(_IZ#Rtq"Q&Jx,b A$R R҄,Wb $p\aXG2@ ٻ f;>N U0һ*jOwKi*^i1_f(VE]SeY*7mE J9'Xes @ %&0v|# 'sz[y8ZbnoYZ ڄc * $LK|q$-3֢E=z0P23 LTҔHBi&8WqNsHbBߖ'9dY8*0kfub:} 6[(Vi3Zfܡ6saN=8R] #Bvh!4H5UI$Ӡ(,rIpJ $)M 5GE16%e(;aGB-<.p'J%1+=r !O!%Z^yqp_)Eho/tPb}|re9DHިUoJR j=ߖ`ZCp';T/=t% ;܎FapBՆX6tW]Q"P"CjsaR˲{&A#;'d/*̱&bqYK\^js"OkL&Q2}:I4M_J(z͗W=A-':~i`9}z P3Q,s>TC愰?$#N x]++N px+;[oìܳ@7>:Fͅ#aa<~}pw-^?OjtfǖDŽD/)!sL PQaH=)Bq&LB1 DIU~5 *ũJPL %E*hB96 SIB s 08Phk| P\/sz0*efD4+S`,QEVIXA)e@9yYTZev_f[Pv!K)-bc Qdi@(M$09L"@~2CaF(cfq&qD{/ P&(1€JB+`ďQz7$yp=Q^0Wo;#W4:?&5c#¿c\T;ote}."{`DR6ğ]_擣h)ɠ٨\ 0++r,di C4eHP "s&` tXj?8+mx렌9cۆx }͟F˧|3nvmJ)'Ჱw/q?دhD+7q/)YEaD͊b^`y6PyU&A/+|Ue:e\Gh*tk<wFXߖ6X)rDp0}fiiЭαK(:P3MNYI|Éy0} q`2V8]2 pN\K)m ) 9t/K?*E4qHFE]Jx^@n}gh2_:H:6+lAHs o,8eriIxu{#iH)bb8iMXgͷk+3/XQ6ֵj8xvG4?Pׯ(B.qloz$N#[X^29߮'^I~` M+B8jR"ov;[I0T/|{Yy+4/,.ތ[yI: t*46ЊR^R2j)&>[:X,6"^hw4޼~ML8Rt;σ68Vޛ^]I$ giّRj6L8:wq{!S6Y2~˵w_xޝ!>/݋2l-co W ᭇ Ęz#]-9EvT ;uG=ejCVmWԶ{tjӁr|ͧ~r:mx?gWs=mK\BD$e4yD(Tdi&78f?oJ~@c,&bv LEfSfw!cKmx_iG.3hѾǫ'P%fAC/}`| C" W^n Y.me6MgjlavaBnPٺ?(0MvMܜ)%srKvM0G%rad1@Nj1zHuHF'+b ¯mG>UCcb`~u3\ [ z +}MkZʋRs\TڞBNkk/l&D)woۮl\L7rN0EwtIqc#S9q6ˠL h6qܺ<=|"̫DfH ^J%cntۢ.ʐ]bwⓀ1z!)C)W-E#HcV$Q(QT80hf^P ư`]Lyp7wSEsl(-2_(4_ Xs6 |@e.n%#TxvہV Qr{I&r)b}9ƕJLӚS ԒmW{UiW.1y6ThW3ų lFlE>zskKvqv2C֒!%b6?Tc?] pES@To!m**Nq\e,O!Lun)+!&Z@bMQn!ڴZZT7Ϝn:qq 3Rvƒ4 ϳ/El\Eƕlyll[Re-TyNW)ɶZ/9QzՇb׋n0PsB!gIr4syR $9 ha0 ۉ|m Y9Dz3Hs슲!82/=!01AWԝ`2T]>ZV!f@KF0,̊s+ w)qAۭm;WFfݖpxa͒Hcdnv7sm߽A"D ?ARF*\s7wq&Rj<0&C`a3Q!H&Rk=(Qp^@9E&s}6Q]d_F?v|kk0lqT3QtzV;Eѭ|\RUVE߄ o-eG:nee\aR mrnkItilG'^x>uԂ6>sU`)8 C}^+H!-1?LW=BY}˜T&jҽ V|, Puq2rgAְA-;p hvjY:d9 ܨZ`hM '$@]A+"* HH32E,2!UauuU@7K*MGJF3Oi)Tms4ra!b\[.2vPpm|x" !Tt1Q׃G## C6$ eڥsPSM>ʺ@3I7qj+TxJl\ 'jgjڲK]cB%a+KޱӐ\&\H/)" ]L ՋvmV\8/$Ll*bvcIw'Axr{ftAV|Bt{ " ą^xzs]4&+LYE>qsa9%.v֞#bXSvCQ45WcZB`'dӵJ2 -![3:jSd*A]&[ֱp:x0eէ|NMtTȐ[MQ[h;ESRnjm/Yӫb$FKwL٠^rӾ&\E;O=U.Rr赧]J=V"xL<"~D4DS9ߖoxp [E2Nh/YA $h٘io&ZXv0E[kwhyv\\vC :Z.q]JO+յEbHyWU p_2-F*EPFuVU |Rg;+}D皉\9,cql!t39.no5AuJHv6w(Wj12}z|E6Ŏ3딖ݲl=}yYqpSg)%G^g \8B[[W$eiL_e~a"y.PP^]rKVv>| vkAߛZujmcXObQϷ_Q[)OkF Lo{fH6ꚿvkrQӣ/x!b'sZa΃NcTouFӁd gh8S#h˭s:4I?#f=,-wdXL&=6&iN`KZ Ԯ}74Ds,qR:H:uOOg}kA%Ǔ%91z+MO&NШwgt{7 :hEnԇ~xt4H뫫__}8x ;M~ cj'Ϙן~ysg|7ܝ]~39~gҌRqc2hEfaZU+KʹmzO<4ad~>e'C-ub\d _c9 9xճDݞO:ŗ@s`ŭf,@X>aCݎsH3NZE><?K4`_;Z>p)uN:Ԟy)'pd,K0~y!atvر£l ?WpR$$31U+4N7C-\/>0_98[/n__ -Orx:3Z`yz: 2:R7<ϯD7Q+t&Hqju_LqZA,h6~by?&JeSH Y6~ev iN&z~[-JރP^,zFcA7CÑ|`?';<΂]T1:~/QV2Wx$1uIA%ÉĔŻ+49<@C]ȡ¸aOаĸ5jeVP[98)gvsD5OC7fu+ҚZ(DB-2b[tCK*q3/ЮH33ÇvЮH-UXa^r|sb!{-|l~T Fxr|[ޏK i"R^  Uq=뜭]_eU!K'Ap8B^VQb!}ݳEAqdt[[pVT]_" DXɵ0͇<&zWS<,) "HUՂ$n]!4,{e9p+V(hѕ;݈*RBT>Wi-Ty+f) [2OF^lu#WBI](|x }F,<!<)%#>״~zȷ`p /DceF0l,p+7Bq0B%*۝? l 3x1ÜcyP;UN uqvh-A'~> Fyt+wr.'T'l}%M}0n?þZ!I乧$Pf}yr;N3gXm]ba6$|vKMɔUSyMn& dCp5؋MA&g'f{kr6=xhoc?jjQmawd΋%M`(͝،?$mcetq`wnK>?'&>jd<ىwo_41#c>2MI2?<7n·i[}a!1}8:ĎoL߷n;hRZ!7Kꍙ~J*BLgo1.x-(A)y E1(CPN sT#=⢐CtTR.dnvL( }+T齧OG\5VHX):Z ')GG+6]xg.,QLq(Ҧ {! BhΔ;g= AVA4,a*~ML1;n\}eV` վ~XtIg%>͘O#ރz B|J.VQf js0.d\f]qyBc_ %dc!xFl'Qꖟ\by6iж[qӤ}zFz\JEbD.hK]PyTBF1ZP6@#- [sp)NF0wTڃsBOrA9gwikB*hY08&3A0p"B]" j7@Rg]-w()$Z!Ѿ<)WFi\悭b\`⋠ZL%S%~p8=&9$a9=*2st"8Y*ӎxgCG0Ž&p? iĂu.NTx{u=z)V,$\Kry1({Ew l;="<:nWz&͹=X'WC1^%=!Z06q#鿢ۡxGa]rIŗ/RaD) % Cy!a*8x/@;B[dˆ__| I\PQ |<mV]47(>t-RNԦ]0d@348T!icFs^@t,x;I7DnaAB3hU@ 7ڙ\VqZR {>)F~yzm]$J@Ƣ ҅¡(qgW9~nj/GL;rB$sYN7j]q<~I袒"(#a ͹!WbІ^93+\@'2(\=P&k9Gi/=t҅8~s]QJ`B CgN~t"o gLțS$af_4ps)\=0\ r;ᝈTʴf=#F5sU `^wINx?ɐC-09PDfot.y.7>nl{Uvܡ?yU04\ArL0Iyct-y5*EK LUݚz% %wUŵ! TI7Zy4LkJ&,_>;Ea9cm9Lee~4-S:= Zz9 2MXXtCi3IMbn>_,ųF51Y3`͔*f/J WSMvugo^*moxL\֯X~3M[ h U:s%+ʱncȳ 18¥# #$dBc % 0LxA .!s\]'F\q:J!m[CXWN(e`D]w/Ճk; ^h)k&t+̮&HX:|A\+?(// 퇠|nF8}@֐M{J!l;B58*Kƃϝ}*hJ&-@j^,k_ 7"x0"VW!٧_~k7PI-g}W}@ {{G.'ky%jXpynJ?є}Y˫X(oBj~uߏ(J^* WGџz 'іKub~Vw^QGpT,;Ƚ8wH2hAHc|^ sBQq+#e-2͒,lIǰ$3Om;z|+V_Oά{gfxh?3еy!U*]dVH%Վ}a|IX뚹`D_xuz&H[G89FH ȩMrY+LxH/J@<VR\ygȃ=!mR@F갱7N#*Zo+=WxQlV u/hʜ}("oò(|Zޟ.O]1P9q}mF9,)}̄fUʪ@;>2 ?yt[_ RD;hcYM4LhvBB~"zL>o|rG01_1yDjȷto:+~&7TDQ2EX@9Mx2^+/zW-YTNKa`欠6p4h $6IEz}BEbtxגN)Y'3r尉z{u>>}t#=t\;H:VJ$H0'z-eђ^FKzYu+UۆYO0 9 cMAoΓk &N޾/z`6jzm 3jm 52V@v^T3+M %N\RvPboT3 d;7}V'Λ]~o'#n)T\ /[.k7gxroiBƙw }`LCr"ޟgxwPšÍZL*x.zsl\u.nwʿGSz_uNx|.^__W}IgKWnN&N5v.ާCn%m.fY 5)e- OuT:WEȕsTQ;?t4ácH.:)sLnqd37@- V9#6qN$ x!eD."xf1Vcno۠՞|c l)vQREQ1Mng)TȦCok&H&kDkjd "k\Sl )ᨭ:#-V\eMQZ2p\!Ռsф}ȬnYm"zy0jň%SrhYs@}Fr4eY lTz۠-3P\lB" %:w#Z>QyZv ͯmQ T$FBn2;U]}ٻM@1%;|)r&_ʏk0io20Oxg;w"ܸ8Ɠ>]f{WS[ĺ$!\Z{yZ[r25m>IP9=^`:]9õjr9UjU. rQ#>wJOpJOgc^rӱkr ^dg-0srr3tggc mfS=Rs[&FG%^b] sdA9VoӿZ.ʭU,7WQ Jf g\1F 7XSF[fQL] AF@t_-{(=ccHG۷[fL<*8݁{`]7;N&=委3Ak\A+lPa"RyA}{X0 6AɁ {''r'pP} ARHN5+:%)L!¹In5-l!V(ՄT3.cN:S邛a wbbN^f`e.d2Hv3BC_N[Hwj^8yOjk뾙tf 3SX2R ) H`R+ $,W&H _`rZ1R%)6ܳhJl͠3fN͠,l+$n6+  f /{; ]MJSfDqB-/tލAjʥ<|M:"~ۤ"/]*\Yh y[ۯ3t@{[C_ r r8;23wH,z9#m8^UT Z+;vNŦU^0p:XbALNa^mPé)C&޶3#Q?tt-JEM$-`)Uڏe>uֆ}Ԁ)i75`'lǼՃ/,GB9)#̼7}J{;Nr &yOrV=T^1yR7TsNvqBbXjCz䭒+ivQ4wznsqUeyYޥLp8>}*H XB;h 4ׁ0!,XW8-La8qL R8QbF޳5Z_ۗ1&@r,F n z9h8-/)ߓ;U=s;dLzxXtliLSJR{j4`47 ]ei*|F9<23G)X廳RFؒ]EQ2#߲嫯-m K/>]~A7:L_,7Ү g'}@@rQ}X]m\+e9`%s2c 7Uz /|û.ry~= -cE5[׋zq`|7} \ͯA1#-҇c~H5zLՏ 53xK?FoȨ(g|"@rGzL0 zcJ^gC9/q&K$c?=]_ۛ<,^nUۘl܏lE ;FwZ!t)o] 6~Qq|y_l-psE8vT 4J`q.ZW󼸽Ɵ wG#TCXA 43M Vd;_xHԵCBM܏{rˀCC6{LȽ/]$fny^(36OE/.SI)bMAryn:"%y$I߰,=;G`j$v=aY؛,k0Կ=DJC ʳ򜜍 U#g9,4_#D$:Q1#L SX0BztֶbԪ8.<bÝ)v9 "/8*3Υ˴fޕ5q$︡FpV5"Nc^ j@Phh4EXeV^eJɕGon>b53Gj(nhgew}ܑ%p{6<٤AZtr探]zIzL`elKŴb9 ߚnD8v"^xhNkamԾ՝hTH}ۓfxo흽rv9-]̦M'4*Qz%PɈqC% \rnZ?جw?+ iٕ剢}<-mYŐCefL^2 Vh\B zƐlfpD# ɣL7Y F>d=v:Lwp~\ُ38j11~ޥ>g˃ =?xZ"qDpÂ|e?9Vf3hLezeg8aǥV,  F)bx?6,Z񞫞wx=&hsR7 fp34Ss0+=q9=!zr@`s=3ښx%`E'f<@uO 1qۆ'Wsl`HK=C$bnÈV  %c:Nʮj)yjv?=6΄f%.g;klhD+&(͜6LD<_Gѡ*݆ +\f+n{Í'qقBi<}cLtsPߚ|%J4ĶRck 㚨 |nFZ|gƗs!͕TPER/܆STT w8hz9vaeit"l'%|}>u~}_M71S"*4+KKDT0l0(c(4y,'|2޽E:»i+3KC4Xv6Y.BƳb;?|E>HxKU?n  kS0J'4d1w=\.tqܫ2tomI&!U+26wvc˧ "цvaiۛϿt &d7syëӬn'ߞ$=_~^=?7WX']W_Pj۠1^Nyx#$ 9(  S6€ ޽b\&hGLG8fQR$`GFN H P* -bkbS:'n<=@EIQ{v2 l ܻIg ɚsc~o0<nQbX"6XYDA*9Jϙr!,`"q0zb]. &|5YK/ p Fj&>"-X!)K(M@]a@bzk$e_i|W.Oх'ک[?Op6'2Im%/+-ADݼ#U 8l7w͝~V?siE,'h7c/˗?]\ AGXy(:{% 2T W˔6brVm70VK7[`)ZɒnJF `Kj`1L$ڀa(J L' t/">|qG<#KP(ڡޞϮ8<)Ɋi]Mg/rWryk%B|TĒH2+F+`c7VZ,UcȚLYR]3e^ET),#F[WH yu #L+mY'\E݅KYъgwoUE&jbQv F[3}A MIZ8O oH8[SMƵE*D0!~7"3?:1LmO5cڎjcvׁ6@y < XͨDPHlO5ĖX{ H IqA7|U GcDEDB+ڈi(Za@"7aJJ r{Ԏw%piH&iK4 yc dK$oZ1EG"*q3+y+țĈuphpukiRU*N^4xl]AUڤ|"Br"1\ 6)qlT[%c)ȓM|j0$:t3e1h(]8 `̷-|uJNI%7}wRNH۩C5|ZνErSEMG71 MIVd"`%gSW"8ʕQ*JF=)4FR" "ͻ{CxjfKCWM3(^ϵtb8%}돭C9+Db4W?(ȓRrtyjA^+UzRR Wt*3~(^{)>@L-#dAR;JUZ}R5&d\ޜxaG՞`L#uybf!<(g鈳6MRgm@HٿeZWlʬl- *::.,?wAFP{M/8`a$"Et bV8x>I V7ԑ@S-HRà;罡jx\xD3u(xLڅϕ.\$o C:?b)YA- U`~or6wnv{csYmш`9/r!qURplpfǻiX:lCPU>4SAt b8PZ$8D`CSıkmP%\j9Hlss] Qw< *WKRtb)0-En,lNwAv97e 軜]~{wrZ$߮8}(?~v>z 0o n߶5I;N Bp9 Q;&f*LET]cR!WfDY7*]\Xk.Ir:2u+^8+f]bN7렾hw1HVTXus MV; džIFWN"-){F[Fz"+ww??\8JtVO>}+v$!O\D;iy{R+ﰃ}](qjs>$v.]l%1hAvR?4Ub5hMiA{%+^֡D˰dnz3O<g(f۽O(egA'&vk;leI}dW5{s xQ*i!Bgy.2AǐڅQ.7lQ\~P Pqs8%Ҳ|yLZw  ?mv 6G3yq9/~^;ųO:<m2XrędZc';-$lR +4(K'+Ild}7P\tC)d J#K}xXsO{e#D;.;1DYEI0#G$EHE|q kļC1gMߌX{ռ J(ܕ~{u9.+s헃ur & \76m#鿢/w{U}/8Χu9H4m@ꅒH $rJ 6n i{5LӜ/oz|5fڱA6#gjMt 6≮܆i'GATxù@;t d%D|߱ !c l#HǑڍ=%@8)Dw~ثpjC;½yddߚH~_2hԺ,=1A@9#T?ԲJ[t@lbK/V/8Tң?J#XL,OY?T^Uэ| uKY =lՐvݐ 8F҃c c?SjɤduٔWO #).+o r[h+t1VHjMHW߽i7^#"FVl6ً/O+o_(B1I@mF5HpŒQ *ueRKynP3&Ƶ$bbWĨ嫓FjqQJ9@Ə8 چK@Mo/dAvzQsx qO]i g j`)&J'%hͶhyV~Jy:!W݁8!y:E!!nWpJ!Gn n~܅0鑛B] r) U+eSNt8APXIXjdYi4!&`&{Eu/ ;G!R)4T}FcQmf@q&$F)Qu2b2L5sTm~0l3 $:>yfD=%/_sF98?/WOsr;hA^1;toD.,"_<^<^ XpT30>fM E;6ao_2A!D8f^ַg1H$oOsr3ɴe;V՞1u8,1CqΙnj!)⦃a!ym9>&݅:ƆJ/O佣;Y|N?I~1N)SPCpFXjzR@aN5ňtDvTZj{~/Dy91vtS,4]1ͫ}˿mXGty\iZܳb%Ą,Td n /<,"wsw"nϱ$w鲤ʂ^ړMM!@$Gb~l5{{)͕W:BQ@t PSv:VdlSF%ɡPonK ͅ&$V**KR"X !1KDbXZ$I!3$0;XP:Җ.&BQjF9qZ@Ʃ_BrqJ;sŨ0y Q{)eDGߕ1~ qdv[fRb㲊syWlDECQs*| + s)ڳ# u.MT=GPDctjVDha{llfjKfL4tnUR@T0pG;cLËz9+;Y[Ɖ7~[Q5غ`*c:׶ 5}!bQ1\g9<ȍ pOM&zW1n^*ZFՇ4t2(feW<ݻe@A4Z:= 0hXPקԹgGj5.iÑz#S|=|10q!Ak%e2dA_>,OA7=hE8hMO=ڞmyr&,?Z_+6h+uҷfJbV_Hv r!pI@VÀUC+pԶgdaO:bg|̗&DU:Y̾&ˤV424*R; Y*<9-JM'Ɛ!,߿:`)uViwipPg; #&Я+}K4eE2 iWO/b-`뾢Λ <ӣm~@[ic  ʝsGtz!!)ח[܏| iԆZ Y(Hw( (z= Ω-+'(.-&!I =WGGE KwTa/'-|T:ü(aٜÅw `9NZ8Ngz  á%gCQyJA) O)cXy~P.F]VA N=z2௤:KyoAQ Pc lCA įkAgݜvuZ`,Df v^SՇge!(׃h'O>\O7h`~=b>$ z|T,IO:[+P_}}@3`|b,nʐ0jn -w{KS|]q*gcq"C~2m'NҼJvn>J&C\amv-VᅲFGdO5OM*>ѻڶ^Ҵ(% IOGfz"#c]0; ֮&Q3Jՠgjܸ_a2_TI]Ŋ\/e $%Mr%߯1!yDYe{CLFFwլ 4"{$ֳ^fм:`G-{\UvQ|,9cxǰҼ ~훙S2팇W*Ic.G$+UIN`T}8`#{UM0*:#1XmYٳKdi/ʌqs> +Tq+ͱUT`6aL:!0>8MОV)mbR c tc5g[IcJ<Fj`6:h ;-dQHRFC'UJV"0t~G Do_}mT9 Q sq(Lq)̂<2NH`{BݥZpl[oR.mn<36Zn-onL0|L|O"d|e4z0OAX`A)̼{V l^ҡs~z]OJvZ{IşJel@X 4H##s%]LxPSR.2Rb2_-##yVً{|?ŮJwQZ;V&Ow4TI[+eiDoYҿ[|) 2v#cep:xk5-]R'lE$b֧^5V ﴗ\4f@D“zl&%^`%(f&Ts,U UxoFOoҭ⾛t먇c+hAJUOUY{gӋ[ɆAR ߩX1l=ǻW;p~l>Dar#V]зNRTa6$I`ۣgQ8-?ә& (A8]hL=K}2Bj:,>GZ30=saeW:S;@?EE!Wo\>Z rK~%9^l"WM'VEo߽$楆ESu1ѬPu owZAo~x&~cfSXP(a`t\C; fG pB Fg#~pDƏ:3-CFnĊA[ϬhJ,_5.KvCR=n%CKchhERDdyRMD.jy@ݻ6Obƻ9f IEU!qvMcGYb-곴R݃ &5J&>Tu%e2 8IhcWUPEXMTJЩImb^$7(7ǪnL0(tQǟc>V0()sw"n _>S\ʊPEgeʮV\._zU*MAy5&r5=-ҝi&(TU@^ ڹC튟++2_8Uui*7%&W-)-QJg2H9Lk#F0df0,ܒ@0KtH"(ʅN,(.rOM5gj? `!/"{eB,?@¹QB>X<=zɄ1r>|i7|,b@fϟߒo>Y&($!.\Pi~&H$& 37#?Y,W|8 ʾJ!rJ5a/_ݿ sShj8=2\P%h7kfl0VWxP7aInF_~z=evU"3wIj2L0)L2 ;b -Y Wa6Kp B^M հ}.n9L 0 *OĺL5W']&'BE'L{^ CcB_jφk s vdEgLɀ0tQ}K5(1z5t.a1d%Į+P 碽EHc۝+P?wϒ{Ȃ˽}}č5Ҡ@ CĄ\D facaџQ?v1If_R(˄d˹R/Z[Km u&fS&3wd72(XJ0 5Ly-FZށP~dk9wA;in4Y.s7_?>Od_[7 [r:$}ҤSw~|]]]]7Yq|z`2J0LJ@PPdpj4yؔB\'xu?xpS,_i,\](NL۲I0" ύcX5/Zo0BA. TF4Ho "okXY4[;x;T=(4߫.dosꝙ=¸V5cX5"Vn}MPu4'Z")0ς=_*r>_ 8oY5&$6ihrhOY` "Ĩ:hmŌnSݮUGP,;<M-0;ua]/I9śu_Of+q/Gϊt_??lxAP@yL57de_еu-4z-A+[7Dx U+ ׉J'dРk=+=di!)m,YFfF\>іver}Cfܬ)ٗ/f"N!ue^ܞr},Vl uI~?Dci" >MRDFT189]Jic{nŝ nE=|Z=pXUr9 8yu[o?ƶ穷6yqgTV7s ],60e%ۏOGoqM-r4+Cc`x1=Oeb%am:ȿ{gd9gu.a7Vqƭ2c$H_-kYXJEVɈt`Jq.u7 g5m0ByP@ݻu aVr9O>boe&LmNAZB9+ SI`UurWđ1 af^?Ӫ260#)fsPe,H'KJ3hyT{1ukQ4XXdnŰ R(pxi+jqo?:ߏz! TZj%O.c_~zeu?U/K>&:/CHe ]dRLb^ګwӋ |8dPJ< tmZMz XiOpW]E8~EwܥB1}M$*zOd )$t]$ڀՆLXwtj̚B"`62Y2zAPȠ'9k{H:jeWZӰ,~}ȳٌ=g=?'H8zQtXRUS~{qq6#g_NqdK).J931?Ŝ_vW^O:?ުSK3z1W$b3:)6'}9A;sOw=^@s{&!`>.RNϙɟ.O.fԿ1j_m.ҼGwWָccmdžΡr?>QH%1{ß;l W" 4x #JZ)/LK B$}ioX)2=VLZ(m0Ɇ6XOo6yjƼXfR#?Fi~c~_JMxoXj,/>YugP[=gNNgkt^fwk,O^Y:˦ՐJ;dWuz~:s<&)`xLIc_\\s?X!)f!7X I>_zލA[eݨ~vQv.RӸuU^f}ے WGd^MIÝ_!,('ktoh&Y~dN`$y4٥$ᴫgm)QFϼB c4 &XAl2^ĵ'cuɱ>b.oQHXq*7jpi!BA!ɝ1E/H-YA\E uGd1'RkM\1S:`u H|Jr9'ce0HW7 QN޳eR[̘9R-0co7P\َ=E;c#mKv kSkombJ8x7O2 f'QOjٯ*bJNʣ:sBHNzC!GT$  nQwXƹDb&տ_!S3L64^Nmk,9Vuq?DX~gi[b強H89u"FnTqk:c&Q:qw+9y4j)ԽH&xQmq" _9v~ZɕyN=9 ߍ:j%sT6-UsHKBKlj +W$>kRרn(+REz٠: AP2H~(u#86Ɯd Z6Z-Q̭]CM'2MKV.Ag] ]Uܽ<(#u+Le Tp>֫;ί.XY~ ?| ̿&y̷~*C ZiɌw1AI`YQc>AC0NA1h bR_$UӜ5Lv.u&,s2#Dxzd H"Eqy(Ձɚ3ķ0>H< U B,9'[ڄķ0ŒL|%# ҙc灳k;oa`[ѹd#XZݘ ľ̡CxGoFF|Zi%XF qi8th:G@x~⭩X=GF8MxTbrIaѲBC&۶e{NW.˳.Ec->9gZɇ6 Z~ul >b|nB؜Eq`em5{G:/f_ʠJz;?yV*q<&w1s.qN%ه MzԚݩd,ڭd痮EdMKִd']K|zqI?;zh%ͭhO]aҊDɊ:攽[Exvc<1iGUrstYsdW6vO9@8%&>$4;jzPJ85~ (l0p^@FJKp^eRCr֠Xtv,s-]ﬡ姟Z@M/|lOdsTgSu~-x?7[~+6Xz"'(ٯsq^i"_OglTe3_ky1;,=&_7{.,Ie X(GN;&cUj0$z+K,BAe֟1@V2@{\0uE2gjeoEyRfg'-%/^7^N9G0͋rvyqXtv6=^l_nk3 Oӫ.%כ#]?|V3u<=cPj`2iZŔ孞+놘Tc:$;*fk[א,ڴxw0bNǓtjW\7+IHOUxYkD5 u;Y ֎x;1R5j:*q|BdoMϾb}"8Ȟ]ѥ \?(N2zdq>Cdx#~"s`&q`jHu,<ƌ yM݅niXbSڨ@FK |P Zi0©>:`bQ1;UN%֧Hꢝ˖ YG&%~~%ĕuߟכ v4‰Af5D~3:h0$g'-lAZ6R39PRi̒EKV{;&Cm]4:y: {r'#Q"YXa&1pL 2CV}\?*).IǎEꤒlQ҅Zqr⭠OsБu&љ 3.8+(}JFv2FM@\©Zn5dz]=!>T .9o'=#I] )`iwQ_pnjQ2LHeR{' .03^M(]r`Lݖw|+<͕RNK9G1Aܹ2OFI:pY6{[AܣCd#r"Hٴ[ۧk#໅l)e]ZX\I&,YĂ:ێ1ؒCp))(Z-ZBG5QMv[J,#ӒߙCE=iʞsga|Sr9@^\I}_Vd:$JrZ4-Wa@K1P2 R:'ڪ (7 QV\ FW@[j*75YwӠԺ׍ZNviNݑ>tmi֗R4hwun͠0Qo}s@dӃ ;mq@zIYyN93S6mMGír0FKw^8dSsا?=D~S7HC!)Z+8k xv{FY=`B^ZVuxyJ8*9w/Ae ܤf}+-z" [Qbh >KTyjJ25N:2P#qoNOӬUY)=ρ4da[ ^]% -ߩ&Mw^-P'Ql/{'LI7lbv#2"fr?b-8>:JqgR1 .+Qr(PպR^JX?SV1.r QjhZJVbNǓ*v!#`/5'R Tl ߃2a2:A(y=# `K Rc!["k1g0y;,J c $"&Pl`=d#W΍^ؼmfY?aӾ֚mtU-P碒)TPslN$;?h}|b*>_{ea0@O-#K58ǻ"de^ {{ec@g^?is0n; nޑN)J?`ZnW%U#,"Iaed`O:YI֏*_^'-^\Z_鬈[TU(qyuQO]n /{߶9MFvW( D%8oaf˄gr<^m/S6cMΘ ݼmn~uVf;)QZ 7gX0x{ńn}}y?˷%Sj> pȎj*ibk=۬ xˣaoχ5@WO\w|M"Qkh5xڥ_@*P#^W@R^(- &6`gzf ?Ub獔l~Q=AZ^NcEngd]k9>2n=KHԳW̶W;`&@ /r%N(X\"dNFbpInãQ=C3ܚ3".5Gx;?uյ!|< B>d` Ӎն~|!XmS5-уjwOmPizG݊Mn0j.) f8Qńofw4n/89 # hWGCR|g=ј5֘qL=U<ղl|AMK|( eOg;J1tپV'ҲCX\Krv~!h (VIQw&5g]:ԤQS⫧u~n̼.ecJK$;VWtK+T#Ub:@}A6&OZ+6FGh5*dƆH; é㦳Bt.ֶ7yA]<gmF8ɑ紮ʢ WdՎ؉:jSLEj!xu$,Yy46}V6T!x8ΰGs V 46|2RN?Il6QE/T{E@GE>ƳA&dA,L1YC^IQٜDtXIc,) sNY ˥ZYTҚ؀<)赍y՞Pálx.6 3K4B*Əw|6}ڸdrv:^N2t8^ @M2 SF"\]bHl E :ywV3IϠ6G0-,SU[Țwk.%%l<؞mSBRfPBFGo\9kohI:M((QEB%E%/@%tDѰPx`|'՞}aTtj^ޣCb"s=lٮkabkM46զ } [ ^\\Ň-nSHN"-e0D%y~ f< rvs(JpRv@ZUp@"),bU ؙO[u{vwF޶lp1V-](ƫaƞHpںe` WSfԜ@ 2v'tCA$yiD;Q}vLؓ-ls1DL{JHǮ|bqۍ Y`AlG/;qp Ax@X{݉pDι&XM|5M(⯕A|K.q$-KtDNjsy?9+5.w`O1;){fx5wD=#m= h=Z\߅v ~;~bia17wqɽ1O,W [uέZ?}&t=ܛ[&cPn{V53U1#hZϤVn.wTq1xǑ1̸CNWBaS(]`K Fp.L1Wg|xvF\IJMݜ9w ;/8lOWEErq<9{߶Y0vf~5vRo.yxڸGk/Yj0Wm.Aot ~g7nVW ଵ /˳p^Sz]kog|KJC)>݈ry$=4r4Uuyujm\=vn ?=|pvd1\8,z;#zTb Q%mVyB>v?C'oG{>c BQpѣ7WVGĊp6ףKG}k->Qx_Xh:5~)WKX_J%J}6TH?a&&vy]B; &v"vąǍeOαכ.\iZSO?ظe4}y?VllNT'ƥV@r쁿>VޭBIXQC!U Ee 5IUe2b6eS ۉӄ-dR1f ʱ2(U)R*%Αw%L&8ɂ.) !Qp:[|o7yW'Mc &>'d< ;-hG N (׋IRLX$L9 LHYeؒ&WE^g(UchJ:+aCWM$v5&'{gGlgjx֧* n.Dbǹ[2+>z_UӨ`~Is 30("}zWv]]\܈N,b7p o5=MeNo|Giq6n?͸A/HMV3^PƫvBПDpT3b`ŎwwN\KCX::DG 5_U7rc~G遞}Q6;uu1'{N00S2s:F'cB\o&[=ib)q0O0lgG;ia% DY$b!RTj}9@T/ B*ǜJ[HsŹ=S K0[ϊmt4dgv[ 7#",nw]X^5C &i]:^{2&N8YKq{.ӄQGDl0J?vϊmz\=Jo @U 퐶l; 'MFL9U U*Nӗ=#츴PP̨K}Ymԧӛ_5z߶UiA׺"͑P>՛3`иO1u}(݈-#AzC,]m(XT-.C'ѤҌ2w*XH[4o2UNPW"n($#3W~r%$B njUVi16H14J&̒RJƂ؟%Ԍ 峔$l/Fh^k~T׿Q-Ne.,Iy!][8:^b{!h|~\''}>)COu߿_DY, ;Ԙ\+]{ثߎ>Q)J@-U$,iT)^t"r-Ozg&y8#IIXs(axF֥%+WY E vdk`rb͙Cfa~҉D)U<M"Dʼn 9d"y cG3?o~*Ob'A%jF뽌ݘ&KG4 E O,24p$%Z8<9օ(v 7!v6Ns)OE;W疷Ww >Fhқ hHW^JZOcɡ:}H"SU^ ۿKGEݝRܑdâ+V]}==Õǵpqr&+gܝrxMj _o{ЗӕO2OԌjZKl; Z_[?9 &kUT\eO=I96w_ 5iFụs մt'2kdAHH}ZLJ{MҐ  RpP )4V1 %%R@4jD l|9JvP(A= z ՑCN\ϢaM {6hxt;C1,%s6DcOQ1o]h?`cIujh9ZJEK-VxA/tHt7/($E)h)eTZPfv m -$MΖZС*J *43 G9ދXǕ{Q 0Vin[8x}gc0r,}ê!EggR9wv]0Lڟd5lRܜ{\B ֚yem<=1%p8X9By$ G#eK"wb} A'HZ -?qWxe osWx{NJmctmnQT7pJ;|zʧpTɉ ˰~G2" ^~Ԭ̽!A=A ?/8AfTv?w 6"' oɆ4%&Zi;q )myMY[C T(_jfc_6Iy}>W-sgk%EQ߁﷠6>P到z+u>|z}#={֯ӂ7U=0o*WRk4HIʴ8Aot/o¼DhUI5QQm]51o!Nwg ڽtvVsR!,j\zW.]dnކڭ+ rD;hy_ݺfj>$䕋h'YS8a[WNw4n$=_\ϯps+%҂"OP2ӾӨYokgqU(|!eFצǭ} {}#3k"80hOCK,5@>g>|<8aeYmᤊEPFPD{z@^IQ*Hǀ,>\eJ`CY&;+*))ALtf F`Wf߻h4, |7UJk. I*B8QIlBR91 UZДa&AxG%#e$ ͥv%@F;?ߪ E]2lVȫCiX|},By{]kP|vɭ.>Em3&OtZSDls7T <p9f,cܜcqOw D F-t0>qROP3 ; 7i܆0Jt=QO!Bwѣja4P^E-˕,/HRAʁ^'QځuYױ4EaKjuŒ%uYԪ:1Ec0 ;RYqlմtadLjyݧ9Yu)F1&0-3 _fp%?Ġ4z '.GųYpyբORڋ?-D9xFa埋N,)yl 9 >J=︻mOEp 1)a01?;bUTquWoB5FwJQ&@}ScZEiYno;q8vqZvQY3}nZsRdrR)7GL&q]$od)9_$(ai z+aeFPAsÄ֏iD'@=1,=Lt!bg!UfINzùV b\k3A eiTFb 5'{cǽrL$`SFDckrD%AOL ]d^&ԺHJiޔ%n.!D~K@ЖxѴ.rjB47ҁ`.#FJ@y+u(SJeNᘩ"]dtUUwӣrrz6.U뗿xEjb*%={׻1D~?tN!L,~~ghgŚM?-&>t6R3]_ 4Wvixd=3xp;3bF7kNq`һzr, "xN^; aBlN)r)քydQ[.U覨 mBF ,nl)JinGc;,K7j~~U~+l+j8B6P02!rhzov#Ae +)e4 *Eb;6CX;&BGi{I76b"0bS5ϐA@T@Q7}֌3J{WzJHO'k<~@P Ai,tHnDm 3@sFPG>烫GˇW,\z ^H,. .2Fa`,wE-sxFOetړ^dIύD3?$fa@sdO@) /'T0o !ar6BO tGäJ٠lTݵ(~i%hነxNf?ǫaqo.x^<FaWMخ. ]|? <0_g2/=p{=%t<؛6m P 0 /ZdOhC]w3r2z 4a!BQ="8]`y6Odg7>tozufsGp+T:u8of3u1rk6o|ۃOmȟ[ӔƦlTx:r78 ؃%g)489󋳮SCyב{`_Rhpgp+2QcBbn4T}v2){-zZx/w]<3,|t=TP?jqq ]P&nZL~|u8a[}^9atv=},.'7x_Z*MNe=m%jz''ǀy-Z2Qji4R'2aUΣ cЀ ŏHRXJ^F⑄VXÛ6T:p]YoG+ vgwЃ!b&Qt# ϥCd1"( Ve e!;՚-妧' Qd5< }UEJA˄Z8LW-"3O ,n"W\HKq C . R:2o_w8! ~^ª&_Id P%)ն6Cԣ4R)|]Nɮ+8QR@oAO!T9lŚ;UʠW6W.==(^.yNK4?sLXR5w/:CD8Ѧ)6.;nnrncrh2|z$0VUu(5cw?5F~gYM?z>p5n"ƥ8{%{i`|n]]^1&Ф j&v:ZWrx|aͳz֋I _`24~I3`ǚɲD^"gVܫXYk\!Vz ;ΑY(M/YhNQbb  ۻlsa N Guk}DFR]Х#L )z"͹ k#a$x,cᦗ1d.Ac w5hLOB˜ #Q!L7!K9}㜾A7( p>w (KrG}0Rm8k^!J +ZXbJszhG)>[#vf½o@*{h"8;Q[:˲alX|ľZ|:'h,p,]񭫮-~bUЭlZt+zx_A*0+VR]G1*paUA˧ʻ nՎnޒNhيT{)j͆zPM9[?*%f2e28zj8FAhJ?*ݟ3P3Y|[%q{s7 rګ72i_-=|@P;օ&74_M2mB7eu5~6.ЃKA>nPRa=˷ql UNwu/8֗%m|,uC[ h/pu'n}yP:}ƺ.*[Vm UNèJYf絝Y^Ns7_Q>1i Żt.W̱yQNr tL2o_|).8L֙ߦ[9Gee<[l8e@|$V  uFЋ:QvJOU&XhkӋd}N3%;huSP>cPD08:pޝ\)<3k|Z*P% eYwͺ|EoCZ ~~8" #f._[T,aw198I(mS+X1`%x`̺|誼5 @`ȣ. #DŽeșYH-et`֮>r8=epZwbtb`k8t Sg nz&hgINzØVBREDg#R[cXH*A"n2PiΣ4v[};SM~TSO%4O H <0c"!$tp$ՃjӓQ[+ap!:UA$ZEZAZ(  E''h~k&Պ7:ї%g}q:{2 ޡaBh/GoY7M; ֓E,ƥ[-nhtJ8gA}Gv(Bs`Gn'Zj݆А/\E7tJcXʶpD嚟BVhk1;DqI\TPVp1^%#%)S ${+SNX j*<IS/ Hp1=RAG^LD& AP)!4qB9 AP=0KJ+"$)pՒw a "x*d)fV9*RLG "EKNJ=oAy& 6CnJ\(CjR.&*43t2{P-Cܠ ҥ*>yKOZ"@}T!YT 6fT33&\}8_(ͫWm>➗&14rn375Q"8FG;-"z'|,F%}|.t?>'oMa[$!E </f[<@D6JvɑC$\|g%$vE!l@=Jx5^)C/ dNN0^ '/E)w!&!'k ɿؔ |dւ}6 g#p9)(ef 4< ik}}qëodD6H0y*;o#N LOY,_$\WIPqcu ѕN1HBuJH];6%cwlF61&]ςlQJHD{^$w?e\sn/NcDRA$&l@j}|W]!51^?vjyo@k^xԄRYAp!*7DcEŨ%HJ:|x' X;01]$Ny}lz1h4Mo~*asz3,e xu>td'(u/lɘqB`%jIBsqF8A;aV[X[X Ѡ7Il5x m&җ!V{uf>Zt(nR }3D >ɚF}6 mLWI__fmd>`c~-KjܸVy\\^ QFz>qX;avVAKѾo}z[_{lzggv#RJ8ncDOnJ5H5|>: [TDƓ `8VyyT"6q '1(&m)/J$?UR$Ӣӵ/GR\p0篋2@O#RXC<QGl.--&E٠06O"Xأ\B *fIոBӌw?i:r+ӂ^b`=/}S!n~pvj z;;P\K>1nL@ (8G>)F7Vg~?+2y]6Okꠣ?EW5zru z p(uS9/NcbũN OHR4(+ U T<3/a [Is1F<`$9jAim 71)R/8VHZ 2,fo&cek}K%Z== w4` wߏ<䮚M_'fGy=?>Am[3܇aU3t3Oy|w)z&ZHTj@@^vv <9$$ݬk+Olx0|2lǠ8AŚ[SP zjwMm.C3 'k\|&6Uoc~LmP`q 2Q|YO8Ӛi٥'TgZV1&%k$r/͘d .#t}V>N5o6Uul:B j9;=6v퍋q1?*8UlbGۅZ~w.cρs`2Ag p7Q0 WJ=9L/B1̋SH#"U8FMT"}*#Q}DNv9=u \sMԡo@CFPl;R׊pS~a_j+O_(s4 s|hRkc9CHOxЭ̲K0w6T>k6mݧ[ͼk=?3>/.ef=%<֩ m~0O5j>/oGޟ_Ƹ{Lw6[S16Iw4؁Ξ[Nn݁r{!h):hN =mк)H A61\dnóP[Χ94{u `݆ d Fu7y٬$- yS`Nf+R]Y>=(@Kp4ӊ7p4:di뵤D,T:qVY؊}), i:JcvmmƖ*}_SmTH.V W 0ں4.J[BZi Bl/]UN+Ȃϭp[Lnw?`Ҋ4zYχҒT,e^>>NlBЙnז!y 1%(} OlrC<$t\}(٭=N e #\$cOv-$Yٻ Q\ܬK@ikI7-N1vAKfN JJ銂~bo: e)TaBxfL'6e>w&)E.F3r^شǏClx|f,asdH kxm٨+4ilՔP($y]4]#4֩(8$c>HBqlJ^Ւu\;tMȖک ڒ#5 kܩXpTOy$(W [>0um7l/z\*% Xۧwaqw+p(k+F&Vmaa'g],1QJ1boaID)6F748efG %%"/?"B`U6Cc띠 Q˺$ E`+b7!o8quV,n(&AК'0q*D5D'{P8iI^(#"#ײ_3Ci,UMle i欗2nZ%dy߳Ԣ %QdJR{mashRFۊ,+m3 Z`Ii:2kH_JڌRuMDoߌt-UWrSz$h]O:S6X1ɬlL]cX\yἓQ9AHj㝌ұ(1ZJY[Ik*OM (NJcYZ?R%a+Jq&jh(1aaHa PR|N}{Zըm~zHZoB>B rCh!жb$ :g#!k=}%MS[{N83D˗Ɍ. &dguQ1 `aUKbwmFӀ?sƷJ-VQ/@n:TS_qfO76;ŗs_^󁰜-[GdЖ5pcA̍>xLʟ!vq ُQw]8S%*;N0>` KJi~dO# bs,Yt׶Ew{@#Ȯa/ƙ Ls~4p_ɋjT]{zyJڡ'wŤ| oE]kÕFj Y:(mȡjLMǀx:/9rK:FcA]! 阈k^<]U؈=wGd1~ϻX8rH]@i.lE=@e@̂_,Zo|0J9FVL*}W/.!N:s&Ej| +j9HdDj['T(%kmkQ*eF¼ 5jY5^qhD&%ct&&44VA>STh+Y ae׃6nxmS[`]g]զؖ{~ؚk돏؎3|a#?7?ܹͯqqn3 tDѵb?Ysqw!MccAOӟ+X9ceyˎCC?Zb}~vSځ0Z*.D :?\/?TdRR20 'x20Evb'+o(%WYx#Ai|Fq*5DDs&$ P)wr@Wb0Wi(Arb$* :r`!|Ў́@ wej qπbOKmZx3 @?1F#l`PXB[Ry8b$28|#M"Nu(: QJclM m 3:H1 )$s#߁G/ՌXq A|É#Q4HئGDL# cTE, m:-hQXv!<ϖz$7ӱ~9H'~%=Q0 y N,ŚId9,3=z$lYE qXSFlY!|l˲%FMF</Ŋqq3ID V}/dp;f6VXL9r !%܊AjM^`K")Wq4eU)#)J)Wl hu赩6e#s-Q@ꤘq6ig?p?ӎU8[hL!DR0ʟȴQg\0͙[IMyiX$f7a6=RD˜K`kNByZ $!Nn- 6L6t K@(fPHF';l6C]gWHy;CT^%yw=K`*}0/O]wá십Iqj*Wg"/"c`N`_C.%_WL0܃nXLV}rPS0% _<.?!`I*b{&ccEbwE0h bks}yQ}[laôީ47_SΙIYw:TA%w1|ڥڭOusr&\e4gjvT_7)s!y]w:aZ;Gf {]K-^kg*}76-K}JXq%˧t2Kl<3 rnK |rͶ R.%=ߙX5x*TKғ8H0;5;|@V& X^sXΕLFwWcք*Ybb=T8гujp@5uDOY!0X|YRI:7Nz.U-?DiTСM&j&:.=O g``9暙IeXKV8k)FnpՍ!ȰVdYRa80pTjV4kwB\~O4H5 K0旪+ cPmc.%*#TY. T,J[UA]W 7*fEJXSq%HW~ eT*.< Z*Cb9E|7նkQ.ݯ.E^J;m.ugտou6:rѬ'dz^xel#ϦC|sg7ۇrko׷?|+ }˵7lmoJ|om3{D7?4@$:̔'n[_m3v\77m̧GtK*{ R?k{Q dȝ~Zwc6P B6ܯNd+oo~xjjKNj"kks*Y:l-JoRgʚ_QˮǦ 0[]X;m vVI7/@T:@TMg%@;TO "w{a2py-Ӻ)s /3Й]#5z`4=|wڥz g ܁*,۽^6xF 0PȭT*kxN D `s ^<@ H9B=9WD6y ,BY 5S>gJW"` 4z|$A,HVVq3ABۊ e$D7Q^A\X~mɘ!#k>gd6߽-f6jvy?P~32bn?L1;d.Yyј3ov)9LH~4Ocw|ɯjUBsz<@GpJAZ6(uˑ׾Lxs~1%;&k+BG/Z89m{!!a| G^i#-,(;qasZG`#*68 ԍj):9Xs =5KABIk֑OE_Zs]j{?KyxaXG85(7XA:Gi`BA|$ icĴ.uٳ{:֑|mTA UGT qȶ\ļ5mޫo՘)̋>5*3z$'jOX6ۂKnD6I\]qcO`o2omK?,+p2]4,&${Q0<kl;X4\ʝl|cc:k]$W"urxMxP Fmv*U]B*'mW؆ϷoOv̇ɗ/EK4*VlՓ'5RڜuV7MO\-T)*Y"d^Z4XU`Bp*OiS _>+brUzvzv3(")ك،-4_L 4/&+̅ Y,i+}̬wfR7Y *QϰQMztΦ63-kʘdFUD}9ֲZVޭDj1 2ˇ9H3HB_6qԾ*lMv80١|!Eye4)WI'^9j}[KQA# rtoȫOKt@$ ,Ihdڱ2\SFPrjb-51v(d'#dï ,&\3#L$Ti!(I4F%)1Jht\Fwz_L> |\B~33}ӽdoUlw!p_njB^?ޮ!t|8=OMO~irwbQ[7foz`:/zO{ԑ;mv[0S&' le#!'[-.xMlxX`_VcR,A*?蒽{rHAQSs}%zd{0JV-HϦv&óuUQ O[$I̛\̢$Mb6"8)hc-澞Ϋ-"mQ*aQ$T"\(3Kw:ϫsB d-W; ٭]'DDYsGbj5ʤсoS,*UZQAٽOtatF Cv;l†XK+Ca6I0ce!4ZC&}Pe(c bc=wR OSIL)E&EN}o9?ؓ36|@g}5)"Uݪ.ċ;lF'DI+'4I,։VC4!ZόSNƨZ_(/81w )J;YUٹ+Gē 7XA BzC^bKo2b#]:zYO5.K1z=[:' #)j/&b)x&M=CRA_ŎSR&+ϮnO% 9pc9kaCUBϧhµ(A0&EյM1O?K c̕,3W*CFV(x:yf`/2D) 0" q"FP$p*ûg5on}F]~{G Y﬋5|>ys{ FX3Ѩ75D摍n5b\lG4_wU .҆yܼ2wmw7+P:Xv h"HNݹݾ#nEAqhu3xY`Kb2"D h*argN 2}pRo|Sb$5'S,y7 쏲]o2C]b,#PA4Nf#IXB&-!b)PL5SPmcdԆKjmE.~Wg|-[ pL 3* b&77?p|y`h4g?3c%B'!VT|ﳹ1xƂدC)y<ٺY-ө]13wzirr-eqм٣<yf2i>Jˏ9{_DIaO9zs ޞ' !pж*qTCXυ> -ڋNRXf S+MoʅmD" P$5Z)R`h$CRm K1KLXX!,* =Q]T?RpDRaKO)}ai-g YPQGN8"6x q k9﨟8)ej>ېDRaXE(fDʙdن65yCpT#eܮZa qR)%$H)ch-[kKLb%4ʊ1 ojEcș8S-əT4֎Վ b!4LP@4HSҜ(%&MwyM<& KjcT?R Fj@<7JT-yX@uIm{&IYQX?b?;܌TK !F-[BZPCd9I!|~wZU)%@J Pc| Ip#mZ =*8]<|(8v}\}[s?5A#IaH ;wZ1mW,V6byu6n'"h\b@)=~r]3VP :Fa_K(%./sILVO8dYD쁲O @RJkWDVUaB"0f>Cfb:qu%86OZҖ^NoTCG:5 ;І;KkeB]3%+( |minYz^+ŰFʧW/%^rBBizƏՆuZz$y=́nyg%E}@Қ*!ҥi0V"JIzЄSpT곫)\b@}[c)i@/ 9}FZJ ~kF&Gv/}՘C ^H_%ژ&?'W;^PW 2 ܎^nMu,g5 }gv$MA! AzgPWV Rzn Qޙް! 6 Viޭp1Tw?_ bW(!ӈ0!'!DU/S^uJ$ rJ+B!ibk܇֥AsX>if!|FC+}(r98,S4) %Gxrd-FE:Z=k9nϰ倬.!65+,^֛(QX\jSRR'G8Ĕc8ӹ(l1=N.vm:;J K#UBWm/qąƕ0\Ŕ1t |2;zm|Pv¨_f r/{"$jLntcߥ@([,g-#ގ $M#5Q4б f8˱W9ZZ2-"m@yY =q(_nSRXW,""G^)B\S/T_;PDhCM;&TiSM xOKIlXv>d͐Q 9@]cOZO\n_+`!2;x @Mʱ36j TLaA7.4{s9ǜ$"ӾX#>ra2XT(-weqH0;- ГyٵcG&RO5θ&Y|i\z yuS-[p;+nMbYٟbaz8{*:,"3wQgO.SyD45o5N߹F''{Y I[wҪоq6d5R lo?4SڌWTݞrY`yj X>Sf.*rjx.*3jC_LzBG~,g +'|8 /GrvYs /襜p+D!E_qﵳF,m!b+Yw7AAKu^])*SYK z}߷ZNl )=wWŲ"꩙gT绫̰*&oЍDcGG5z8o31]>LJ9~ޔ!  `sa@KT&"o+.6F`*+P:j>'P%L{<)%U$9/7;IK)fIquF1IT_nRmozR*)OJec)򤴦PN[JAI)³xR՗TӓyK)eOԥ$OBJ)į\IJZJ󤔹:<穾ܤm$'=2MDyP4cSuB2(5eزmمѢA5_#, 37rQ;,BiIϠkI֖Xii#\ 5ZR|`4t ?]RHo劙3Ň#9VAXv/fو !@U(Fʠ0: #"j !M!Vk/hFf, -.^SUIeQR\UBK/Q,AJ-5 ʍ@:zō[]Z G`ҥ,TU`J5}NWEMFWTKiSCL;sS08 )Daх i0Nj8±w?zg98^RLNjX^%5URCܤ%$5BTFǻQ7KV^ÉӻҎMj輺NkwrRxю`f–kт`)\C'5t'aRŒD03ge6ȔOjșagRC\\ZaBT_nQ:pޏy])^FcJ)<)&)=o)L)E9h90`-U2m$wEeС”NN_&\`i&ՙҲ~8j{{2eB{RJvJM56˸NRzR&OJNn&SJIiM54 ORz~R͑4nNZ|1lu6}Rd˟׉lw--]{K76@OWIK'Jqv)տ/*V 2\_t>@, C룛݄݅u!螹:Ԓyq%NS;૫infu-39Af/B+Nρ=B)f2Z3=5z nzt5tLDC`x}MaQwe8#A(ϡ|^nz 2+ьcOג}Fim:{nzvfES9'7%ImՒQ92x/Qe{Yaa*%kc;/%Piz GQ&?) אliOIzؒ=WTVRX؀VWV*v  1.zZ=z?|5ߔzp@ƲTMSq_V sW/@5~Os~ ?> !)~XW| P32K|Eu_A͇]S" =ŧI sЌ2xsqj|oR%1pAd|+&Εk)kȭx" ev̛S+44†~j]՜DHM0WzԗfM#e5H\?T\p2PIJR,R'^r; @_k$NdiJ摧Brs9 O>-5q8+*_DDNʾe^k֬Q&wQӊʊ ֌!!ZP?ZGf !D1i ȳBP(.;Umol"bŽwu͙0,G>qz\Z1myu]'|/> Ǜ?<"!x`k[ȿ.]mU~& 'Y W~"TwΨ8Ki3]p) e3*o~qԾov LGh=`(]ԆN%́PEd2DXK+B]Zgqs_>o+t>\nvGX牽$V7zhVUK%c#Pms @q }'`,nz߭RxEeo?]^2çqyK:FBfN  (e1u 5 ;Xc0p38y_ ,lYYfAHz (1T9tdb@BM9CphFôx`8f:[aB ٪St1s 8H$N:eKs!X%"G65B!YBWI.T5+}_ tUvіyCAdo`Y{ +7ՅYp+A0&Lק U%/QT<f眩[lL6\ѯ GBO^V5+ B92 ѡ_-6L6y6^dGVj_aAJ/zJw12<4sH؍ .%,#9PqAY{ ./\l)uJmٯ hhvg8176z[fNQ14u K~uY0 X|fQR"ڸ#Bw"t5D ذ!ӧS ptĊoo.B[Vˇozq&ZBC_m'n>w}qC~ְg.6$Bu+PQ"rJ! i7QܕۇkL?q+wr~ȇE&(n{;t?)cӢ%%;?(S[ey^S0>΀ Bؤ_'3>&|љaZH{[ MY,g.5;r{Nw̉awrEJfQ2w$!o\DdP ; ;[(|D't*֗ܪѴv DnuH KY'[(|D't*휧X۴v OnuHђLyF#>@Lj.w|@vhCFb+/ZmWsR'>Do۫iʸNIғr @xVpdpsaG{szzuތ0rNr]fޑ֩]>a;3rޣvx:uZ a{^ьܟ%}'yNp<7/fjsOjZ nv6}S /4_h- ;r`vN%V.޹ӫ|ۻUTþ9uVT@cϖ [t!@S/ aADwDAuo΄nbZG@JЅ' .l]!|]B.ǨptwDCu%J)iˉÐ=%"̏z:/:A._TJ] X+PQuP!Dgw7U-"j?'QHQwsh1S[n96yڼonK:raZ$W!UfU\A. (İvp ^foZ?"Fwf9a/'ZGn3t8AH 7*֘\54kHU6?)jH]qeJJ!eF(b-bb+gTd&AS(&nrzñ7MҼ\"tvA6r' BӰv(Uv!h1p20$:˺o(jRD[[ d*PIwhꆒ4dK(`FDbn:2zegcf[w`[K,?3|cyAI f~bӑHxdct6l_WW nvmʗnV!S`d!!#@N!6jC2J{EdƟe N }3Uz?!tgf6L3l2?"(:~mx U"#@5&DH(fA1N9ׅB=,\5cu"|X {{P>>[Cwi"!͉lr +Bxo VE{{6%M]G7Z\B!מPDcE.3na猰 d(4f0YXPt&;riV ),0b(k>kHBX3@[ 6' $Ἔ [P(Y ^hN"R(u٘X9#8@IAEftQm}<t? OZv9=ɵ&`ŀGu1Mғ @3~{@G /UEy"5{_?(Pb7ؚ]~}VV+U ʛh*?sPQۋ{pAZ,]eO. =w@ Y:#V/oݒ3Zu–6.ess7נMb0 -FrfGfZ.\;wry_ l84FƊS`"*R xC,aTHXdž*"{r!p fV8* ʴ,=uKFb"x4A3\XZaj]ê0 o^Ǻ+n]wqHRX3GlǔP"ًeQ,оZBX {u 1\KZ{WV ƘCGl}J,ۑTº ϵFƟMnջBf~5#ZH`?TJ֮T+K%ؗRSǟN]D:j L5…>W6lTl :Z{VZBNL3gSGv:{ZM7,*'fd9Jv.sc| ,߽˵sdi}^L:FvS@“c<[E$So癞cNnGF9YN{17Z=sAQSx_}V0Z }nnVcoM^\ڒe>\q:SWjDz>33"$+,&E₳L-. *؅>ي ^3ų7-/EOi 밤w:|>|?e3M%$VPI90;9 9} ; 3d  #\'LdY'/Ed0%;M;2.๡LB in}nɚ.]Y{g8!9ӣ2s%Ӈ'{SV~PcgNdb5г]B)@n>Yj KXW6q1 SeE u$D:11<= A>ˬoҖFᚵI j77\)bv~= $*Oc|dҧAHX3;Nb{DHi ℃ A%h#CRAWؓx]^K7q9T qaiV4d)F'DjcO. je0:Ei;L/e9"L$kE6v. #^Wۮ^Z}=咒xLϖɞ^YU2dTBj@Dm.*vԓ9t$~ ;NK%&B^OEƅݏoLy< ]*B}^f鋣E8x?vNlb)dMbĐq!! cl1L <&鴬{W&VJLWe~KYdOsO5|gXqAbN)_j-ы!Hi<Pf"VrrjaXd]C~o٣/D5~#0} VsmjdV|{[J3AיX Um Lc`ڙ)$PH;զHR(4ƆǔZrQᔶMcB"NNgc1m'!+\p @l %_?L X6./~9PRWUA"uB?gػָ$ZR xLiv|(]GhOۑÓtc :I.(ri:cԑnӾQ~-?-Q5!8f#8_J7%]P urǨ#.hE0PwH[OD!Jp4Ry)o8P*'"c8\βz=-2J5-Ĝap!#b$"ЫT"*bb$֞#sQk,cǵ_aAYA5)Q:IP ƳF% @U_VICl+}}N'GZbƟP;) %gVYS^!HƥFM0zE1@8miŒc{"%%g3V\2l,( T#5>OCH9j ۺM|AQ ,}XIJ&4%I 4(a?adih<")nW,%vk!xH }|"EaFN*U^M3v*Ȋ,۬ ߐ2)X3#*eԠS|-Hf·9hIo6(.m}(tvt$R8ُDq]XҢ4IK Lx$v2>B<ܢC/JOqu{+$|Smƒ|󉁮/؃WS@{aW+ox-cdg_+'B́pOE@7~h:Н b%[@^諪9 }:*aɔs<4ul_=="&ũ2Vu*@_Ǯ1<]T&\.Ì+G>]L"|uvu 0qxj3_Mދb!{ 2:'Pqĕ!cF2y+NMrq4vUTB_TL`  FHy)9ְĚi@ 5f774S8D vS"F ͅ2D@HB)qY j&i8!"JnK'w?a{#1kt+iYs<8ܜyU'SՉ9ო?vyc#GZ O멝( Cws8Ou?3嘴YA?9q 5U%!~ƛ wQ㭝i,QxmYDmt^]Daʯ|pR 7w'D5!9bj~W  |LCʽAK4-qh%E?kAR-) 0d[g?gQv^9.X۱Uq:̦ٵzpP[rR8\0(/Ȇ'.μ~Kw>ɩ[ѷnOdp4L;VaérMB"nQ:q]!Wq6P~mU'{u9R8K^cv(C|WP?}yaԘ0挚3',',RHIIÊ׆pJ!P@!2!F'NQpHC07ː[?{OhcĈn90RI4[)wڔ/?5^`f̚yPXFJj,^:nf9KDެP %%ft%CCUb`\$Y/B +Gv n')PpD۬jSϧfyafn!OzS& v\,ono! /lo?Z5^|7~#`a'*kx5T}y@%h+|! y!S@x|14\L?D@yP̳𝭠 5#閘rNh "dowxk|>!Tpt-.41y0p %f<̞bcJ~Cj} sqCkyl# tYh\V^W|4n$x"e_iC;*Iu $ZWu&U[$?v}]cdgi_$vW_yMP"HHY~0qpQ:KiRBhG3kIa8a$Ҽ, pLB0.i `9hɩTm դ%jYЂ ,+D SJL%*5Ls r 28=| 8ܚ(~ q,! 2!k{Y㾗Bn_7-RpwIHJ;L2@E]Hd7K$  h=$z) hih}5Bk\R!hTN>RvNkŽ ɃF34hbx"b<<ӝBA&NWrzNUhFKBQSeZVd`֦vjsJ`6iBA>tcǴF(}ոD+g.gqT J[xt*^33kYȏqll:'^D`2@"LѫPRQk EIAׅ%-JCna"F ͅDaFH$b,V3HỘ$ ޣ/bTL'_1Ea}Qx?2 s%Qdy\:w8HlHky$5Ru@ٍΜ"ݺHڔwDڴ_kaGaNfOyw9ydm&&ٻm$鲻t:xTv>*XQL e%f4-hh 52>axm8]688̐yywn:{9]1i 4d :lK"OِEΣQu%D=,h)Y Q%D_!0{K#\ L$oarzXU/T,N"wtVdHxbDȽ7Nriɰ4-Pp.ӣ0˜U)k+W7{T͓xww9p>N9'? %4i~prN9կSNtʩ~]qʫ(J͡@&2YmQS?4E@q)c &RZBBSHDBrkds*ib5SSUdKd%CY'Î;ŧn'&A޾fZw2KZQ{_.hk !@얽ə/`U^N=]u t-[+b$t.vK RzubGR:b~9Az?R/(M~pYENp3[?u3_'mjf/_],-egMȔ)"iW-*w7N|Ofk2 i¨Q͔N` PK,8)Q3qPU&۵K%4,!N$OJeQHhSb%hd~y$}@l)ťfd5جb6tH1o֓īrS z1ܛW͡zaU_: ` 2u[d^l>Ͻf+L0 (+KXd)7.1iF<3(&HU B2D9PW׊࿐mb*h0<͂]a›`wbBnY̜?Ok +^?OWޔ{]:L7aizBB>h,/~+n o;Њ1|ފ]Høsb^9??O TUſH?nO 4?fy4x, g$!,|@z ]a9ǫ`Biz$|\~;PXǣy@ '߫k4h<(/ ^k`6ng3_u$!BM,6E8x]314ad)PqT7fvO(Aպ0/) ʋ[#w&8MQ.M׀ i>*D0zuof1+o Uu(Z9.U?Մwo/rQ6/h|DD: oarhZp5߮BaCB!nS8U̕kZzҴo5q'YS\cpN$t=;vE{SJcoMD^qBz2mggt~Ϯ*t$~VZRB`{C|dX_)BJsBjކ1͵yZv'|ub4I)Ӛ2j8%J :RkB" 4L\nSRJ&X[qkڳ{maclR2NJ /RzRYޜMj_ܤ|6y!Z^p{_CNxag(lczSTB?RPJDO1Kvs>]cs3Jz.TZQ0vCTmS Ks^pC" AJS}jUIH9J)qRh!!ۦ:AHYK)8) xTk.RzR"NJQ>~CTmS6nkHI)F&Dbej_>U۵O8a }ZS+%6kj%3sۄRV"Ǖ4jE*4T:4 ?m? !Jbj+%HTM•&RV P[3N8+@?KP)K<2Ff8&ifF=c[@f%3fc ( ࠑTt8$H:.@`s6ږ\bZp9ڞіFJ) G[NMxu{4&x3F8`!O@1EU{av4%$!r1KIP:99k103Ƴz7+Av2t)U(Fd@Yu%J01&0~O#Lj O껾,;=޳L&'j43՟i<ۉG;tToާ'R[88|v75t}X4@Dqfeᅨ~隻)oUxdS?2ؓ"ϑmE>O7WW!7گG=ßWMu)v֠|X^~zwf hI aPD0|FbEՄ7+VV:+*AhCTmQ^ڞg/E%G (jh`qX+04 8RRҁ(ylZʱmYV|.BrK墼W%gՋI ׍{""7PE?|l?=Z^VH.}=u#?HocҦ <7׍},/&y˶Xi @R.y48S,[˜(,.E~FM6k{L"k5G:AFY9ұ!&HqԺ R }ˌk|T- ƍz1udX_?}t?Ey7󪅨4ϫWޙ IW[^A~w&-K\{7;hŐ:UUC,@`pG0%׍ߌ[W@A@8 nJg[bkLM3:xuj] z)pq[l/43mbwmvJsyh^~$P$EcIthL"#DiFҳgĊ%* ݊-"/uYKed]5s7HBcdWZeFV>/W9׽a@3Oa_$ZHnsײ"ʐIe7M>K(?PF6bgCLel(Woq.'_<+O/b&cϿB}h9B~]KWwEs^]!G9p%uk!o?.-,nkfpRˋlbtlj~կ{|.*x}P),-ZE-cNbq7qWg8S bh$΄"a2B85zlFqxt~?N"x 6o?.;; [Py_~,*wuZό)7~6m۾#ԗgDFL i6Cz?Jo5Ɓ%ƫ% Ԏ?|vWtGLno^lψJЍ?m߻ݫ&僕B3C'i7MY1bG+#JFFTDq1$\,en@%Jj͆ۦ *T\73=VI>~NU]=L=,(?AP/ZY|Ù]yz[gzQ֣n5IX.=?iYv=4&um2'gݍퟮn]${vYӸ4~*3^ԿO{E" Vт"gjY3UU8kVޥW 8-*ܟ3\- "#W`!W ~R,ǘ=(-E--~K7OEKC5Z *$ΝT~#gV*j4U]t{TGpuR=: E]v6FBK)&Y =N%M Bir*ʇxUJ 6 M0e~Cm]rw#WJDnJﰑRb ho=`W~bK\~xOI+O3ML[jhl&<3c?xAeҫ/Gk07 k " *taP e]" 6BOFw^<=JPGDĥ`(~bQZLbՆC›iL_LXl dbd`SbX?8rxYQ_$YsEA:/QtNR^윜Ř VIƌEUT2}5ARt\kرD ;|\dwOx w.7!)(ywo-ؠA[fpZDmpҦ6B83 $7-D8*˳Hw:sΚf\{ߔί/_m\k?}+ݔ~2LYr ZӐ6dHCc[clZ'%Ta5)ɛگ%Uu `EGQ*R7{mo ) 6iFg3 gJTQ5q+  D6-\U[6n[mַ>m,תCؤ.n>i>mBfZ4||OUG?I-to~1L|)2E5 N~קަ\^]oݧtQ׷>/(!V''[;uxd$joԟE6d#@+RC r<~c6x}c FoodLY1$i}逨Z-Vrjɀd5VSw<0Xւ'ǐRc2E`QX80⬕(Bl2OrPQql 7Zs;#ol j."wd*4wyA 7l?WΝyBE?:E`3L#Љ2jS8=}}ջmק;o#\Z&4CAKG*RuǙ/(Q|X.tHdcι޽}'H&RGj8ȩ  g=Z ҃,Wz gb4-QiE={!cܣ$G.ܣ$~";1lIqg39a=TxπlGX|xu)` f=';oJq׈jOf 1o(*TԢeVTŌpHGTJ.5Ak;KSCNjc̉a5KPz5 Llb?| ިL>Z!ڟ;t%rhuAam.RyD Gqb/y̓yQb:Zp$̪ǣDs ES,%̔󲣤I_K\}Bz_4M& uשc= 4 :\] S;WHUe#8]3jm,Wwd+-%#Ox+gXlFߓ1QQݽ5bCF#cd^c*X[ jj_jb9XUaksRԲQh"hJǧspނ*ֶ5HFsa@ƴZə HF[Ur"sȁy,ZďZ4 oFעȜxp])#YIOڅPZfCs9uP<~6=n&G=|,%É)J[27ѥ ga2F*MHwߎ:*f.HҏR{V*$9D34%Ha6fwX&c;mgF̏;4'zhT&P0˩<ECfOFj- lTbL,}Ob^!x5t P "~8*t&_-yCfS}~-s{"v_} Nw^]}ojs}C)=;5m.U.e~sMtA;n _ w}]^p6%ٗe(FywNɻK wN@5rs6uiАg;: IƁ;z&f#4YT\ Y4d@evEl,Uvl<'`2f2fSJ#1 s$fߗʀ|}vogՐ}K,QG>^VA$ }һ\J, /K(/!SIur#ӿ0]\6koG0g9HbMԨ}o^ 8??~`ᘂ* -lU SyMQ6S# 4UE@ V(x"/΄bReXZs4a^Z9+!,<ׂdg ]r_$(~֙򏸏՛@@6'(8k<B(Swz d5. CM `-6]C]#9׳,Bs|i&!" ٖG=<E8r5#p<=1EDn^˄OA]$ϻRҝ44&<)A2(:,1kbJ.}OB eYA;G%JVR%q*>oȇk9 M5: >ojFsTŦ8*X=tV?x;q{L<܃̵LXB 5dlTܒ%uU U簑FO ~fiJa[ac8Sz E6p53E*R5dr5# Ӊ^-`2kjbF8RC26~X[sWSPjTA}kvuck,oQ ֪Ӿ[Y",Tk5w=u 1(e$<Ոyf@ń W +ߋ"*@rB2F")fkrcuYsP.p{ (*~Kn҅s~\n’||c۫xc :, c۫IrRѧJJ:ګ`5$;bgK<xXN6`)']lDvXB80eLլb$ʚ *Dޓf(Ek(՘~U|;bB|OLq6enF9ќн[הl⾒r  Jp\h}zUMnfxe'f/ h5{+3F6y}MLBB_ꦗIó%[HˀslY6O.dl QEH=duܾBIqd(ÌV3V%[ ܫ'e"]ɘ`.pJ" S^^ Z>Cd>ECPpC+Ej MA/ā٤@Ö`j.syleզƦ"АgY:%h;uSܬ FurߑbݖLg+@iАgY:ջ\ǹg݀nwV){fivF w\+Tl\ڇ E z*k3)!8&_>yNax3y}fYX 7N+?U齣U`d+ץ7ZV zAIkB%52}{PV^]V%هp;왅5*i<(U%(1JzUI}T"!FIqmN^)lmx.hW?j(a"ɾnw_pxiW-/\oDnoR3ۯ2d,k ~s i W0ۨw\~&~ЏLm[4ߠxIc[T?UD. &N=E'ݥ, _P_'g +UNUlQOR޼o7FBKvKH[˖Wۋ\IZ y28Ptqi}hj-#հ"+?3- #=8 %e͚g.%xeyҙNcNYR1`~^g_ 8.1yKU!jhAz\{˕PWg*Gh*[K} s]u e1B!-!ml3ugvHcb7O()^A۳94YTm 81E j:=Oq2*Ӻ_NYۘs ҂ 䰅OaD99ce II甁HFI}]w/DzfH,9?\%B-N:3lK8*K HK,lp;mⵓZ6m>+TmazU$ߊ;.U߁D_0MN Y4h'7ɋТýlr c8[amO ,/0_•39\=Bg<6a~3^uWTL1wR|K!5 Gu,5KP@;tYfo8^he] )(h܏>,(,GW`|`!t䇋ٓp1n?OtoFv){1C8[6^]&`'^7l7?Kw?\V- ^>Rw-Ew;VmsAA-,rVOa*dJF*\q:[I[0eV*Ln8|kT_RLVzVJ3>QX)SKyR,⼤=;;7q fΩ%o յij-N͊`eoDف`@ w`Ў]֎l18RNF>ъqĽ%llWcveT8i*^c**A6P1Vfv\ un0CU垲:NGyvxvќ L<]u&柭 C-Q`_u|d$1J B3AZ4 `#atRM1DVGB)aȢ^>" |P+[m]ypB-NX'Ֆ{ (D#1H*MQ6 tN=n:wl+4Mi8W Ce$Y\v9+1{J;!JiQg1'߇k6ueE.UR2vbqyn:`[tqi^2viTj`?w~:WT ިx>:+*R2fȾ䕷VHJ VW_~ X+v'=ڊ9U *T+j #Zzm֦by.379CF*; ӈ3t 2Lj)JOJE~! lgJSOm, e2 sAk|uZ6gڇ2 M ^LPL,Kl$p0K^k|>s5h4`W,҂\ xMw![f(e\䓽1}9֒w>'#]ggy[ӽЯՃq6{*oϿҜ.̕_՟_4.UBf6>\^Y1Q5X3߿x2WwZS^M}~7ﹱJ﹟Kǻt}&pm#2yC|"-}$ mP] j?Ikef֡в ~OdjCZKSt\LX(X]&"*g ! Y۲vyQ{˭:@t"W;wPj&9kWsfyZśrՖ_Zo娵2nu{):]wM'ÖI|ڋmTU{W_S^uBJ=oPPJb̝ S3"ܙp^#Z Viuapnw7ְ|M!-ruU38=9M~{TfU2]P:+t?} fk \PӨj6u᝗ 9ك]p+F?g٣x2:m>>,dǛ Z8Jw?N'*$t3L'pd0w>Hq|5 2:JsM4h %J(£Pa$X> BڧTx=I{*(=~+Qأ²^rZ=O*o~4B,wJo惸5 np HIJXWVPgz_Ţ5dV40%qYbl ط{lNf%[neiA0KfX,yV֩1_fgC &E2R6YFd(_ϲ$.p~&l o= ~~o䣓>Fj))uaѴ$:Yf)9ad%F#g)IƽRF'UfaTB@;Iz 8Q֕\۝cc83EZ;V 30ޢ?/WX.ۋo׳;r-\{NCωuH&ԯ3B444fl06Xc~Ak dl/MvC{,QF,FQDxq;wn誼ppz(uyi8,twhxnto0IQc)4cf}S{nH֭ y*Y42wW`Zr <HJv%}Tjcx.;>Y6* zZʈVT3 މkgy.MKU9L<I}K3SPP*ASZ %@c7*%ȼA5,kZ[Z:DejS}^[&R(TxATkuNYKU1-u9o1dI5 ZzZ OK GtMﴵϛvͩt`-"NvUI1mZ?}V2MDX"2K,f;se1I ITyx" S;en׾U[J@՜)+o\/,\afoza'7iӕe|2J=o!!؅7~->߯nD^*K'˛meĕrd.^f&gLQPE2)|agnlwDawQE/W_X9-f*q:_կPJI L\+iڮL2R#}c$֏<z.<~14>NZJiYm,3כ%NvYMSV An(\<۩LSsJ`5$4>dr*,]ڡgwT&t{2bLII5'qҞ4y:ԻNOe kz%PJ]z|&kv~;U%O+*ݵ9K3H.Mz*؊>zzFl|?h?s*gh " &I#J$5OuDtZ$ @D|n8)ԠDIax02yڢDՔIhI[oaV0մ'] T $hH$8y\&X |2DkAP FkTn uO_~%Ɉgo)O1,q I"- %4I/vpo3\&Q.%,n2b6LWl:Ɇ+J@*bބ $%H  ~ᣓ>]qHi\R۷"%RQ` U,R*(`9Bnl=7\a 6ƋZBkQ2;uTHJIJЅLYA%O%J CS2b2jBY)?MUjDzG“bNBP&_l#ۮ^vg(F9Do8?ӸzpMF/8An p (v= &6S*7=qUl.A[Ca/T9izR9Ѻ@xV\+Nʓ`5ר=W2O!RLvWP ReAYZÕ>o[?*> #H'zÝs,R#q1p{>%k)9f:*U~Rnp 鷀g 4y=Jձcq\kY徹h3aMa%j㷮u=jBwW?ܙ~సwZ1⽯ 9e8 lwvm6j< 9љ!7ie]{V)+"`b<u#= N\,M:#1q1$Бrβ bCts^6'68«ȣͨ$qF H8Ȍ(tiBBkԅTf/ꂩ",*SJㅱYn3hnCIIp&v-35e:ؘ”f3<12dPXR)y!)a*5hZDg Sw`S艷 0ƿֻyfWbc CﲛR\gvӜ(JZV2#Rew Kp~Tw]{W3#[ߧU:c,X2VŗRe_=} >~_(M ,>f 6LK'?gO߼6źp}m}c7Q(ͻ?3a.{7ksncOWwާ栕of^G.7}sI$WbbT›{j 281'sN]woKXpx+?m\ݛ^N!ăc@7+fKǫ\>> 'ag-g@TJ{@N];bᇺ螛#DG]*4(r k2ҠuR̞ygiirP0 j@iKB&=}wZӆldm.HEe3`bfYt RXÜ u|x6ڟXOj3%j"ͶWLρ#-/`"GYN#އ!kQ: Tؕqŋ'6~;զ`ᶂЇfrb.t( VtxMwUy?=ކH#E~dFᑴʲRl0YBiYD!^*& ddSS]ğ;.WYІ>}R+} @}bECxC63Z:@D5}۝j K0:;a-lRi :#uMZ^6hO*]sNu~nɋ>\kH˷um̓&eˀ|;{ڨ>c Fs^G%_d@;w5W>(W=i+Wꌇ1ηkD-p>u偏ĎbݦfLkC[9OևrҩG8wr&n4Bs #L*hT* (9 &A we }uuXc{]F;ۮ'Y+8sډHʾ;ϞoY鳰 @Jխ";ZݩuGSmmw[BJ0 *XG4輍T 9wC5!"" S %| "ZT"9*t 8EkXpVSF #V϶'tYOz+8|:@+CptaR#Mb /*=&huY)Ә Z8/֗{W'i%tSD'YBA GU:[P zM$u+ׂ8=ث\\:a2 A!\9n+su}+$doQI=qE嗻KmIInowul[R[$k#K$THHdw)r |^kIzUKڢ J:zcls Zх0+EcX=$6x!rG"xme\I\KoC!Zv[@-*-i{4#HJKKïx}oOSg߄۴(]__m>_:Nj[ۚPL3V4q^r2 BRPNZ/IE|둅c{EnumQfVY4ٳ=NoiLnFP!̬Vۘ޴ַFtrC;EC̝Bg'.-AF#yFu x_f gWf^3F1fçJ^U dR_$1ˤ&(5p$d(Z t.$x’2ְ>MD ^X饊ed7PaN91rK01)XiS'wCV<Ћ Ȑ0?]c_3_F0x| +\ﻻ->ߦuoriI.?;WB4=sՙ> MN Q{#+5 (c}I&:Taڢ @۴LF76NFWvZE Pl#IQHhƝoHI. =:BhP9Vplte+Ւ=iK|{b6{f1&aK3!Q+sdo$6:)k564'kSpAW `e`mJh~JSahI İWQGE0V# ?-X^catK{Z!o Dyt2)wã IU/&J꣠5AgoTњl}:=ҚJiM&?+ñOcnc1(1ߨnNͻ,nSXDGԻi: ͱoxe=Y7Sz),nm35%]st%oDI\(;n&sl&|݄&}6 XA4/& ~x"fJA%lҹU߿rD-,Z6*C*~OU"{7@J%PU_1GL9УR.(>K2D6?9%n>6A7٢"wYOyE,ޏڪ =/}BX .%IW5S8ğ>n@tTFki>a/W_ %8"]!RdvY~zC7fTsYCvwuwk' J $vLs ޏ)̗rXa&d>(x&)&@G"+fעF46IlW*2>X#&"1 Φm]LJNDh. e $!w1o(:F18F#wIP&93^vf8X4 _YGt9Bqn$gZ)Z[.A -߇3PiM-AXj4'4D31XALnzIv̺B+Y>%\]4'ZgNͥVn'fD53yHuXdÅCXvo"@z8䠸v8.2׸+z p!]u5y^'KEP81-8!5 JUcFgȟ]\_!zэKV_rۻ`toJjnCRē Oצ0P`](NˉAd7%hHjjPB$AJްli2f+)#y['S| NHN9hH^)IMR0-orJhy*)Ǩ ⧦uihWUZ'nrL{fɇHDZWWKrnާUD_m\j/}9[})'mrB9wYuluoGla s!{N5_'TJiCi:g}Nˮ󂟜}o| sTM_\"? <|Wlw6x*c/:|ׁl H ނByRzVnIS4y65ݼLack]JT hvϜ"TZf2;+RꪣlOb6xHJ/*˱˳7eWٗ˜fo83V#yHG ) ,2!OËS8l>\@e!Ҿx4 m>*D#:NlD&+(ӵZYηk ,m ۱TZS0IL.W3VQd5%')RGl.`& +T(FxEJ'֤Uy'tRrFc[!5Gw1`= >AߎRpFԇk^V eCjI5Sc׸`n~T~ CrMDAO|㢕 n-xzf^sC G dezqᤣA:"!qᤛaJ>8DBR.2Ѓ/6^ܮ^qKl$u6:eS^آ.kuiaV6"1jcBa}~n-MUjmh%(;J(K,rBKa&%ZΩlۍZSZ݀ BLjCm_X> nZZ=:yԹ5.K.-u^1IsFQ.K:YjJa#sXq&\@ĊbUۣ68.@bRgn~̓ԢB`f+`oօRgL\#:hmʮ6DBs6:C=87:> %4wFK Z)2&!!.vbaA.d+{sBi pLv?_~[Tx*LlqS͑ǣ ;>{:FJ`Կ ?|h:yiЖ(Vpc=J6EE|M`hT:~w]cr4j#76X7[^熜vA(Y}Jt4LG͎OO)WiK_ld2_A8,^-ɢSi6\dJmZoXZ PkfCsl%r{ (fRu3IG* >E0zrrq)JQ0[L2_8bR|H:Gd7Yl,xN/5,ӣtzL:s(Vkr` I:2:%~>(iNoQy۴9 {49^UMb݉S ,HwGU-ѓoKI5FNwz7/S )aoK\S>+U .3{0 bq "PaRpiBI*L4i4ꂐS2~9y"'gw9,?|qsnMwf|ҿε[w1{v1oǴ3I=]`Gdl"Q^4/oM߅Zf"?'3xvR9.iijex#NHy?r?/uт9Y髶R2+L`QX)`Υ '_ })f$:_t.5XsK_*QfJ$}G>[eV:ZӊڭpW8Ri?d)7k?DZ5Ղ%.)[ғ6>odmi`6ҚM}7[?|F5mM;OׄE~λK&wSviTg}G^ܾjԭkd+(yWu(USa؛q`~c.-))YO9y]_twW~{|_r|u+ fG'V zG89!"qNZѽ J8y: &48yP׬K9yRmh4J]M4"4VªZin=--j: %ɠňlhUεhBθ΅(1_o[VV +5eڋZ-L+´PsFZ|Oى:JX|hBH~<)݆N6NKΰ!dz7pz>x}oOބSg߄TB]>jE/s>%w[ H[Me6Mz;M[2,LQdNB̪=r??QgUJ aH# 8DB]G(AȺ# w@`\bȋa-YҳARTJ[b`fNIQPLuuswC"a'%Oɪ؆yH,dʙ#C҄4@eN}J,T_G7pG&rX:VڛSy- [Ïdߧ{i.L3fςBݫeUvo_]w98ꏮpRX@N2Մ6߄#n-ױRwa"Dud{w8Rs@޲$hI2Np\vRdS) Kރ'؟I2y 7w%?SBK?>J,%x:ScYT =R̕/a֫ooqkr`!l D#}::ŪcrJ'E(zжW<5 /O atLEiXy'G\#ɧpiGƱ+HơK`AaRJV/ń ''v Hd%Kq$'jrNYTy'X2ω g.ƥhJ8,@;Vı G[%~<}!nu>9j5\*q0vc E#@i'2k<ڽG0YoRr\6 :C[xe!"cKN2½fdn0Ц3J?@S:X4ا]XLlCǶO@EmRlu! Nhz3’6=,D21!hTr&}1܍RL_Z& (ܓ)VbfƘ~`JqR;B JAe(p-"B"-M/R>RKp t,* 0ׄB}ۛm02_8Bl#sDf$,bwQ}W\zxqzc͠5+NDSX),K;sf1ba]WhpD2dN>D"^œe(N-~ [0MuWLODf'U({mm|N>!%;,1BZ!q{7MvngfRG| 2ϐ=Ӻ.dྺ&pY"trh@۠M\q8*ҢoqwxJ]hొDbJ*S}lX}5/=SɊ1PjzM]yMrA5&DK]hY!kgYA5 {oo5&hG'>_~8p, +5 /9ho@#A12&'%BJ|04~X@ԠZ?rV߲\^eV0Lv8L:Xi1i/0I1Ût8dwDOøT$82'W?0? ,#DT8Ҁ_ﯻ__-QuXʃqDZ&RJy:R5\f,۩ W0w$}dp S]>SN ̎/ҐҩB@I}JMNhEuQźQX :uNѺ5!qSBNJu偋mu;L4Q‰[펺Z&4/RڛEGPqȝ!SKENz-ɝX"|O%V+ ǤzrEڕɥ-6w]T߭QH o.F.lbjj624_gfV(xrߏfsgejrYOXa~x c8 ,7G/~]`"w#V]dXz?,!7'/ȾA+P&ey)>}=Y>zяBdކ?J[/(SwF|~.~.٭Wj<__@j4EnPzLL/Miϡ﷛Wq3XƏauE4rbIIFOw0GiL_ ~>}t0h. $ gyC?OJ`gR$ 쏇(gzʂe}bϳ?7/lq2d4էo |,_O oOoR'a-,2VoP8/R2]EBΠMwy魠+`޳tX理BqBmAlgfDӰM_5̟oRRwb6b.w(u`17\V_.=ȓ$[Xh1/\{2NcF_V%f* P-UGt{DQ޸C mXc.,Ͷ޿ض(Fp7#BxS9;,3B< 㷋5u5~htR١ ,/ǀ؂qD"))::0g8BV<l5&9kFOn0_^*e+ܔo^u%0;r|5)Igq֚pzP:ջ`dU6mzv4DK$b ՗*JR}JtTLJN5XA]c!ĕTs*抃2SKJT t Pv#Rp#RMPs%%i)gVKi!)twe RENԖ)+FO?~Q̦cObڍAcϦ0ky~~Ycڻ2!ɬxL0Z,PovAU `SMgqV¥0.q{ճWބ+$aw.D eO-%PH-eNEP}N%2&Y`w?R[A*K>HL(wB`ݴ%׷'!盎OkMއ<1W8/ە}8zf魃x<6e(j@Q`R#_+HHxMF\y',5!&0RDϼ-ḿfͥ2NQL%vH['#F`1!l Xi!ZQ'H0Ha3h6~HɉNHFHP<郓K-YGi/=Ε6a TRヿ-p#iG!!NhR22 7J!BD5#FAB[d9# #  ž2(`\!Q?)G T%n:"G'Rh"\RR@*35Z $). qڦs;HQ aV`El@nR$ \c.D*R HٻR*#e.\FDaAaO* $aiV!1b‘Xrp< ]T!v4T8nquzo{B >qg3NuV:Gt39SaWꌌFpIHUC}uTArpUDÌ%rISs.f`%yĿ%K$)'{Ws b:HP-y^E#FkR tɞKG-| a73y|x&#En6—@c畽"nTsD{W*aX[bFSj.TK[-l-[UbE|Sj)nZR%^:Q!=ɐHbC#%%.~ _2JviY-e8S-ƖZa,t;.zlqǵO G8-u̩fM}Ztدaӎp9Ք֖^*G-U*&"ApJFjui7w%b%7yqpv{7th{'BDhBJbsSKb_@'5=iD")rՄyd 䦤[ULP Cxb+E8[S_4;?gBGAmi{Yo6LU/ @QJVDv&!%w!㨃1WM#wg3?-. ˏb6 ; X nGкS} %aL{ )7}n؝L(^{H `>S?M-F]M񿾿^ w-MHTޅ@ϔDĬzf;uBOؔm2es,t)3| ]]fYwX/eeоCIT #%MژlHTlMG΀QcPHeQ`-5=ӌg{WZ~\1b^K)Oz ߟs<P:¦(Ũ/bŔ*:pev!>ߜx"Olj  VCr`Z0)o뷒}۫؊I襟FDX Fk*Xl@[xZch{IbU ik`maĎĈ] *Q&,;#rmxHrb#⵳D񄖯-\V}Arٮ:Si^r}i['NC2)h^e3Qћ>%Px_0ֺgvBZ̪Pk3&;.fx"Ib QƀNguY>:HHm0B[<W6s6< ѢnVz e!#kR葤@H=.i"=$*,CDJyO-7>HZF@ g3JZ9GpH pI 1H$t? MLޅdKT Fd"b*D\XDv:t TO:˛hRIѐnmDO*ze+$ITt  "ɩd9xz{_~Dn"=_Ǭq]gi{ࣖ]l:oV;GW,|Rϔq6:c/v 0DN4GC#e~ Eq_XK4*W] R".zp*d嗺\ta%2en`Kz+)~T pa47 AI$jP1xNF8ziM:!;yn3$Cdh*g$K I:NEowdyfCxXT2M. OM4;" )DX^1* c!Msg6BEr^?v̻@pƟ|{&7ZvMבī(E, A]AK$NiCUcp;2H/v~ZlhI?Wڭ%Vi$2_#l 6(͹zuyZ$iO>)4j57n鶠53[ $oM5P`ְ~vvF'AGH+1pVJHN%4y NdPEo=pNH[1 lA_ DB7faRb h z<%e|[e (h=[ZQoK>\_a_ 0BچE*DC3Sv(P֢ 6]hhf:fj36l%Y;I ^G4BlU"PӤi2Pw˾%~u+-,tPTuah\!(ARL\:_1 UB)%6dה#TEd2eP*3"N-RKxhdVYRynse|qO-/z[%wc̕XI@?C`D\D %1k7˦:wS`YtkpbBNYmz%1dd\AŨ9kBm͎oz!`vji6|Ts9`a_d;&!+՝NT*}fFo;A.JM(i|*pg`>߳/Rqh}".;h&8J'DcJwoPsv |8(ڽx6ǢN}mp=ziS(8~MJ3#ZɂW͵}SK1_0 Zj)!t[N۝5R)k-̑ș+6{ ]+W*Z3Өۡv1fSYIp TL@fASUPDk!@ȵʹJ$V֜Y'] ǂI=17a0yM Lvi8O_Dh%#R5.DxU~y֋06 \.J\kOGJ-ZW dC1 6s^, KV&{RғfYR$R@2ĒXj(j\>RަkڬOhc!$WEYfeZv˽.|~  [T(Z d.+ !IrdL_S6%"$XZU-C](hiFѠO+<|^ }ыU['~X ^KIŰ-zs;= IQ֋~8C_.0 ՠ8Rg_^Օʢ7Ezse_DR*s)}LX .DD(xȋ}Kjm<㠇KOsRXәMU%jIdOزnM4ϬgnRK+փȊ \B/7gbTvwn2c bȸ)y̻@pƟs};'ZRR>k -*\JӰ{CӦѻbRbF-fX\o*S+%8?i&Cm%Qj(? ^ Ut L720pZƢw][oKr+*)yHX`=EKtHY俧ZEQrFzeKhfX_׵,y~X^>ɳܓsG1 >1-ĥxP!gV"eXjxD.4k~߸ pO1#O{SU* ''?,;Rz={K:qP*ڏ^'Pr^X@h$ i BH\`)t @6QiԼqhۍz5++@ZgkJ [uAhUκG)aSY5C!m3}<>lBOɂ+m?\?_kE>(u;d A2z,vvbryLHBr>7BSqgԐE`'f{bD/u)ϏˀvG2M }N cdc,;(2il]K9]+l#%F`2^1t>oZũ|~ީxTgjY]1 ڮynAuEo GhR8sS;|8Цl3rl/9h'RA-lj2HuJ!2^8ž|vI\F~|VZUaYEt T9H&&?B1,⌭F%6m3'eY(qzZ`ީRD,g-܈djbHN,BiU'\))#S.+:=@BFxh9? FZԺ!G`+dQuR$'%V k1NI"?ɌS|T}sMGV3{98`CK6 !L~; E}⼺6*8[`?_^PM6ڋT %gNb1iT,I*Ɖe4c@#GmN+#@2$8 fb($cޙݣK@EaaGQqYŸB,C:3EXk9=Hi`V3;8?,j*Us썂q(yZd_#%FRq0:Yq3D3G ىSOd KőFA`p6Xd0(T=΀{Z)y/f'k77Fw\`;B BW*m׬ ]p={js)$m砏-`"E #h2thy5:X~\P͔!:Y|;+@0rp`MT%|xЅ]&i#QX)u˖ޯu7_ܨ9#oҔc3D;M{.~<X< s7oebޙ1ҏUf U*eFz9wZT^V"*j%lZ+Q("` b۴8]8ϱ9U6[/S  AT3 ` [f4D뉂4N бj :rr#WFdBIXYvo4D"JJ*T)^4sd)ZX<1qV᳘_(JtSq@I*K(FWg~L 0vPA&f"/FHUhbyFj:ʍiU9+8jAN\L>VA DL/%rZ&ٽ F:uK7-_()$eowD:D| :Y!=dszuDhv B4䌘UcmdIXS۸ ,hV\=͌KN4b[`P> t'"`%Q9kT2;4iQ1V aV5b=IC}6Lm`{=G8ȠqL.2SR(@2F 9H c`HУFd|Z|w^}P;NWWmXܤgVmig(r΢X~_^󏿾e[ɦA Y!dJU7*T-r&o`h޻#Ɏ>sk/a󋛶a\rlZ/N$v`pL(mF˥/˰ywDw ќt azҞwld|Dm;sЧ4mϰq`q\smIg.Y%YճV"$Т&EAK--b 1N>@,5=Aj|Uu{ݐ-]T:ad!s<RE*Ga_%T2{P|>y$ɏǂ*H@:lRR5Ir݁$[RFAQSUqX3Hϯ&#w(OHX}я5λayQ_9v4Aby H*, oC ޱu~lG^RZƩ~I_,+bNŴ Z\Ձc;~|ȁ 9 Ex!JNGD68IZĒCӌM2c4ȝӜ|(s0.Wl7Dh WMLTɣw9uBKFrwhD?t]aA]l_ӷU[M0Xkk&Sّ+ QyrF5]>X^ꝴ>Y;Ѻ3g_?BE RkqBّ5nhYU,sx>p@-k^8tMa?8#9Fμ><~\ꜳywr+ gN }>W;}ZѦ'Z|<_?Q7Mo~ە>HU﷢`mMz3qNdMs<ECaJc6Jh(OoHpwb}Gv٣nfggج5F>*eZ_=75ɻ;~"A<{/rۭ0rgTAtN,/K8- qG1We A@E**ŢvKRSFӻ{ʵ.R8kk1Mn'/ cv// ׫qpz嚽i3ڪs -(q>FqMVkrV 2E D9ŧ θxPvX^zI1cqKǵFMYZ@l@+A(TF哬"=s8#c~l(XZog%5V#$M#./K% $P9V,`uP*؃eXΌSMyqG )~mͻj>YPVG죯/`'zWF[\/w0*0rUrfl#f0TySi?3ٕD{slHXeOnng7>v!y4IHK2BDڭ!&)dRW)\U6<̋dlcUgAi)eHkhdϩ?;xq8\>sI#㫌r{ ͭqL "dB+?Sw"ꥣjrT}0Gh^ $d1HH" iwpy0$wN!3}4uL;f{<h6_ԥScDqݪnT݁50Hì\4* VB =lZAXrG4GӤS.Zj0Ϣ: 8ӑa=\=6O(ck>Ư3s8J%ӷUy\; T݈݋HJ@|^M qm;fE\37]OEaTL?n5 P%㠎/69:TI)M*tBSH!&0Zi^P9}Pv/*ܳށA-8#Ƣƃ[D:= kʬfic К1&@;#*gyV/Ze']R }EA:b)yјi(z[p2.wC"[gzۑ_1izyC$] n˒[r俧(oGllkw,:hz֮7o"4gw}qG݊F+ Q;*5edN[/iyEon1.`}2`"6^eX f62%-#"̎!! ~@`E۝kţFڣs#+XyxOpڱ(~X?0z_4R*#]+nfR3^}}~"ӭ^3O JQeIZv"2@p[kV(Wo槭޵ Q5Ry\xR.49CdіE MFs.;e1q=vd\y-,AɊRH ifK/>7zS$}\sK1hmQjbm="4Ա Pd$%x w<2QJK(F/ ۊ^ZH.߭ dр[ѠbܬR' ݻK#pjO1K5ȶMië:s:mn[})X|mPOgO'XhwY淳ܧ%3NC&aJAyp{GjJU8}\s^[wƻQlFȒs}s H3+]aREK/RS=={$vBI[XUqPIK-ފIUQ,Y6*Y9wGڧ zuEkYNa-cv|$]Ƨ0v_c!d]̮e3aG) Hw׬X"gebCξ9{ʹS|:o'ɤtިk6"pb6Ѽ{6}"E؅㓨sZKaЛ1)Y%*^:Ox)nŰh \ި9-GE8$*U:oYamh.*YlE}j⿶B{Ru4[g(ݧ]C! vY{{T2U˞vTX &x`cIBl6eW֑G:՜C:V^l㚻wFe~ CWҼ<ԈJC!3BB^╻ٺ ?:61r#hyW\-IgW;Dۤya %bOV1_j`qʎ "!ŀz;x0FpqwQYǾ5o׭}Semds~8Pih>l=𼸞INpӫp:<%U{Mltg,rdV,o'd0dnRVr*b7 lq[eQϽJ;(l-vFLpel ?/[)rejpL}>p?mccR8}C1A2pwF@7Z"څ9)(Yȉ~ lPYj ]1dg Jn m8_r}/Y0Ո.y {jC.y2X?LkTwB14@=#A*^t֢lw"m2gK%t#U^YgROgs};=jř ɼGkZ-/u|96/z!bY'F-G tB(K/>{|uӢ^zZ_Nn' aI}n+$O\Ab񭻰jr |B30: ٪z\ow&S&r|11`PӰ%e-P ztɎY%.ZXA0Ŭx5KUf {DDVކ[wM#nHުNJM !,:h/fޫ 'Hi8BkzUv/Weki3i[jo'h5<t-MTYφnU]uɝIU 3<v(1huDi+C21Yp伫dKLYnϦ>G${@gN M! }2{0S}tFKzd,δ )]DeϤjT|qią Ij&|J < Y(/6Vj\z1A˞o1A<end,a%G'eʡyf ?7"ׇ܈MXbe88K:rܐd-;Aٔ]l=wkmV RgX$%eжq`#V` -> pe4);T\}wl`E -B Rwb\{ 諟, ͭp}HS:'O"x*ĪK;޳;HCu: ``!$Q8JGP)d+&d-yP@:KAj=03X |(CoE|í{P20"}1 lyܳC֌Wp%+}LdZ|7lǩCTEc;1⍔ZDJ9l1:oŲ1e%[>q23|kmd?^Jdu:y>?k\/J^=lGNVgt88mG4rY '|J2!yYV2AJRֶ)I ~kVYkB4MY)K.y88D_Q_4Bt)~ǻV91V~yMZw3#r? hcP- rr>[Ĥ`)f7M@LV%e^3sd8U=D+3H/„|1"o8ŰUKqS1JQ3'<J1j=Fz*Bu D obh8ə .sdg@ Pa;LHm.hu_HێH~2H* }%5D9G䔘uSVZ ?Mg5,bkCr'0h w>!g*TbQ-8DDm8M$P !2#5ne֒6#J jM< 0Ra3[gpпg~*(7xPW,TB#"8a"/;| ʑsy=-cƥoMt$mJ+M?iE#|/!ez$8&SD ,gFJ̊A4y  V AV4u-߬UVP4~ǟ̧/}yq_b{E>iшOB1Z\b}8δ;%c'_+"Ih/>{7񛐧4Q`x/~tӑNz֣ Q u>巇Wb5[ϩXP迭Rn4'OY_nal Ӛ9wq~3K!N2|M6l IWI_ؗ ЅkM)@^Z}E ϼCmCK^7iM@Ff_kɻ`8Y;+PE!W-<39 َs"Bץ`IsޗZEd NЇ(D!0z,>߼HIt+wʺk4B:{]${6453 XLv Th H8""ZA$HYԒ)\˃wcoHr8\:^@#Xe2(_uIJ ؠ>@Ajl]uǺ֫ *n^7qB:-p#Xq9hdz'YJTդ8]gw][:/ȑ2k䷇dD$)ͮo5K4Z@ &e62_xYɢ ! q[;,AkL W/uLȜpɌbhARR ر@Z=9?i8I7!sNLƅCR$>L趞ƔX0uÞ%1,!lm=ds:˨cfɭ;Z!x{:*v!p5(,/g-c^2+GW QT_ebt&7R1Ck{|Z?1*یek屲L-cNh)XAb?;{9ɍPGn ף IrJtV+RCۤk5 u5ől2)=Fie_nѯL@P4խݍnGU;ya5j(AF鐜z2k[ўY"rPPT d@b˧>1(()eTowd Xs*_Km4Q}(p  )m=j YC`)׶\pZE?Sglc>SM*UVMBGc*Dnl$BPQv7(Z4Ύb9MŧuW kH~ge"D ҳ Q)@MI]|lbW?V]_-c`B @+N[?z|y [ _?X+8]OƠ=ܸ/-93iKk^!z3s7қqG8շ 0$A؂dGТD^K--~91rHZc.3PK"$.= c>,j;w(tPjmp PDVTNgJ!j"˖mx u REx K]0ѺSFSބU1^FN D.1) {J#۴:@V\rEKvvrn8vV:i"TrrIp& ZVn֩XÖǐ\o|>KAH6 rh()S({%T88w q7c{*^^ҸW ^EX^xUNo5M|~iA->$v[xL>ۡȃ"kIʿul0Ww&%Λ 7oŏ_||~%^tʘlkpTA6DQR+G}5qĴ][08q@xk"z 5OWB+rrE<ȄN OLGa$.jLW]0pxY HyAiE"+ZGiE^WGGIvܡ;W׷z" Vbx_oc&ɴ[85N8"ϓ:/t{xH+E+ϋa"Eɷε@d-#O~D2V?lV_1HC,,}wP'.Wc"aEADBr[ͺ@bE~|yo/ѓ aǕ}Mį-\/G~s6jM <xUqՔZO3)OUHzPCJ; :jĩ+K#YJќޤlX*i` Ӝ=9H 9/777_:T&̞ Ṿ 1v]7W_}#OtMwҰXzb^|'Ufuim$6XsUgU9 UφDXtpm_{+B6voavh+q16lwc.U>j-Y5]L3&Fr؇#p:CwNPCKvXZ؁.gӐ?iVN$ӆ`L[VՖhyqs)C_Xm0gN'&y.2gl;i ;qDAA9  cnsts{/d!XTɛR3-(6"ә1łZ+iҫ0B`:_ꝯ2`; U$2d筆\%H_bc Jjg[ɠ8q2r=!KAlDj=> q8n6.N!֒n3Qj>(QZI!F@5.:ɲr*G$S*~1?WṄ6Rx囅+}% -Wюj"NynBKdf''vUtI؃O`KǞo6sc̡qh ,|Ǽ`18U['M ]1Ej%鵉_󈔕U4SH㝶9$3žwdrZN1{^\vx)9֫7 J|)9Er!A) M)*xbQ`(=li{9!y@aU yx Yg7z':ؖX0or37>ڶgwQx9h6jǹ;B̩6D,QX odL9S;Lg8Xc88q:q#Wz:VqTP%8kl4N1m[F %T%dU*%Ko]%|6iOӀt+4G,z ޳o_㌨}VoO1-qRִl eUP,F(]:;7T|j8BPݢ㾟d}%P}ܪvPMdV,e2Y!mf.O 5>s䚅+<1r?d3MDD_-XŗlbRJj@5D(XqKIFd^oǫ f"Um;STGUYWkK-BLaBq"uMxU):`1ד!3HVTaJ,ɇX/T v< },be\ T;ĕX~jI$Q}u{HlB#G8kG/msј+Qבuƥg隹,?qK{4gbY D۟B֩)QRRXVP4EhM.g2mx{DpdZzr0|-Yt7:|w=c YdWDX_gb>QG#|~/Կ"`Ou=/ϗ,rkX}1 UF.iz&K;\U[ wAC?UYjKfo)AZHZPb=$NXH TD_nw֜͟UoVP0_.)bگ2F ;oGp#%ؠ\'PZm8( qf&,T-}5Y#e2(,*%r8IRzӽFeH/l{)EtRrD?:Fb3UI{Rƍׅr!ό8JbO^cL۲ˑr QcNhֹBS`KgڕxtrcI:&W-s9mi푌vX8g;5u8xP/VQLCM8#?;_q\XpPRIG AN 1N "896ǰAS݉ApLGº2l?h7(?s0\ npbw4VFi\1p^ߤQL`qɑrk&`::)##8,JS=5+?qOxqLȯk)%4,XC,~YQV6S Tz÷rNix׆z}&ONJ# ǀ,OW䒶k N٘X?IV,<%JgOXDpd$,=!:b'4OKCc)7Kga,=OoNowĔD,d +A(RQn;f ``U yhm˱ K6^_]TM]LҒm$$WeATpmόm$2a[:\|*ւZyl}bdx$]ـwPG:]囂bž1wQWoEa\ Ye꘱2P(Vȶ&DAd@W-)J =u;9kXhDOhGZ @^ H)w|Z±>yY_{Juі?)vnXڎJ%@[(=N%Lx)BJB8S,)Y9-*LE$[xz%5G]owo}a}s7zٝ^`d#Prꏙm= Yfd~GCM"] 8jN~-njpߨ$&+.ƳΑؼ$)pHj 0EACu2\{iwb6=yMn'e\"ⷫ\dwcA(i{H$@YGcmm uf:HgM'H!~y3\Kp ˙;(=S(OF;ckص# \O%RvfӚed: BZ–%߲13.rd>L*LA"`8u'p}HUVծ(H!ƜZT4f)MJijY.GLVg)5P؃9@eP.V<j "U`egV̮\N鱗ߡ:ǻ*1V5MيRRT?LNjN-+¢[km% Pҷv3QTT5bېٻ<0(8I !PR[! }bnVpU9*xs+sCUUi(Sljyv?X"u~h;utTZ0&VPi7nZBl`IŮ>I&r&~#`YA`Q'TwS 77W/u[.ݸXZ?xS./kuՄ6%C&,A]2aɭu SW*%Sfm]Ǝ#2`gB[YT- :@XvXq|6~[y/0\ɥG9Lsz(޲">Jj1"Nf޳nttQ ̀'G푉|[,v?.n=>%؆j%l5c;oD~urU"U+5ŹegAѣ+)﵎*RR.*k*mA7$hdKuRp=Y{m5H)V|}7"#KON4NSgFHw':C(FCȪ^f0p$Ýz?o#)8I ge'PmJ"KӺ(|")"xA@0r4cn o vbOcǥ#TQ6pXH&jo\GBR # b f|q;=Z8BX:YWUcl`u~=Pw 6V %-U&6^#|R0Ccm.9= fІUDBe"}U -y lS[;CaյHQ[S,T6mWAؽ?0L7qà uxu6[hL=펴] /G&ZP׍^AQJPͨ79FP)ʐVMKh1z*p1ohh5,..ʄPRFPzj- 01$ԶHh &cSU ".ɮU8-J Ը[xKMr&ܤwB}K1̃B`ԆnXF?I]OQw PlRKlBq?i2[T0_jD;HVڤ.I 'S*BXvWģT19A Dblěص09d3+luc_\'?#fV43IF\ imQR5#jEdP3bv`=ϐ}9@iW2ҝN9#uSnjQx~6kq7I=iF5׍}8rhS4 -)Wɻr1teTTuJ IKJȫ" 3]D?@:3+ K7-su]va9ˬ5jRJWXvP] ZŘ`آHfFʁ]f] {Kfð޵jT&.:LC@Notsx.#=,G6HȎoU3uQ*%/ FȤǼ%m? hf(VI鬾ϓ2Bhm<dz-J> 7}8fέq?Y]<7Y$$B4U;(:'I21hnHaU&JgHj&%G; TNCw:5}5lst:JJJ|TIIcA.ݹ}'{C7tovp8}ϭuj=ԈMR"4,mR**VuZ hv!kmpВtb&d# D|ǗjA=_Ga;\ncF)Xضkw6d )g98K&i*ald(Ƣ3*vgޕEd_7FZ,M{qM[>,<>L|Oբ%9s1F6y ph r m܍׀|O}TNA 9{7D/澥#m\F/>sw=ՅBk`lERKuCOѨ`&BVPAXDU+o]?{w_134M KJo^WfQ r"q#K%y"Orp߹^dRu&)_55HJ6тj(")1/V_WLq%gtqΜԤ级d8^ pڏYو-UAe6m)JWȚGNf<ݮ'!@y3Zf~n9jOM |QY];3ތ#4.j(hMA( TRf9ܲ/cyE#z)D^P<}te-yt2apPšF; kNK_@vA ͈ll[QU zF]4-kgJ3 i3B;M5pFm PCV ݁ kT'QZL2%]j'=NK7)9f&R,䘭B`T՞iO 9ܱYvVy{_qi߯n&,R-qP,:+ Q$bClB'/qL_[JiZT/B]"$I* 8\縭W#ͭzVClo҈p/׬Ӽ[8spl'\1):&`&xO}Ywy_1ɒH_ |Ќ|M56Z,~ [V6<</;[OZڪL-&I\I"VW BD%Te `M|$I8&!UeYREOv:'(F/NlB* Ey*v E>JцWL|Gؚ$/o4SA|yB>P&=Ik:(uILJ pTNY)`sۡIKa+1Ǹ Bh[#y8Ώ!g9n&zV0͙ ?\wRD u`jW$nNQlQ O6+ 'j}J LL`VYq|.+'|3u"Ӌ뛏}Kj!'jo >;$#" icTl=whpخmaĦ,^6swLE-qQrlul :_E՟/ PNRKb˧rtONWC1$j<yv,X\m!ȸ,5k˒e'AJEeɗUd gbyXLTaRckB`PB MҘ Xtӹu/Y;6f:-!Aa: t}t@\nS֦z+sZ5:EEyI68 ChMg鬳n; @˷͛@R޺J׍lI-`XU9nN%s):wY{q|ǯmc3\j/qm  +uEll!^J"W2keZ,]x_yG #QΜ͛aߺCPiBR|歄ONxI`}圊j)ۼ/Y)6_ݚk/ [z63QIjO(,V&l֔]76~r ;1:L'Hw-ׯ%wLE.>:񏏹ߏtzqPIk*`rͻ_'L>UPse6JRp1j@RI+SQɢ[GQ&Lշc?w z/BW>wmmJRU,= }Z x=gߢmٖ%u,gmlZcȺh?k7_% pؑs4P_~ fZswtg| J`k@Dͮ6{qf8kdXLԣ\JS%䵩n6Պ4B¾ mt6R*:{&j.;`K[.G^ˋŗ-[A\?`l o*^ʤطBG-Cc[5Eyގd]0tYYF.&3F?Wb :WbjHڄ@j:d00r`SBBJLY~n, ʂgsh{N7OT9w%:X#u^1VCU^CJ$[3纼 GE-&pa1Pj=a=)ëhC٪ }w|_2, r+9PHݺ^?N=&j|;htYZ@8;ĉGɨ^? %~m?K5**e *IrMST}v}7j^Q1N鍹1gyZ8f 8SV\tL0e[Ͳv::G; FtK!$Y7\[q,5(VY fݞ2v6m5U@LGT5ÃѵFlXSWrBKsެj P#>0>hߤڛ n/Sxc:;?kP=#9 }x>Cc5ճqj*@:nh{mghڒA kRDTƉ&%5zܼ"emٙQ"cDzTyW#h7%fkqHaXr,ni֣012)ٜ g$pѬ3c(т0lУ}ot)~(}NˌM^V <5j4F=CFlUWRŋ1X.d auY,XS*Z&T̊JRC뇕v~-IA:e+N(l<+T+ޅE(4V{5A֌hf{DʚG퉘-ctvO :~D]|y;Qgo(/cig po걉_eSmb<;1Bi= V}Z۔ aʺ[$1@n9{q;=|(5zPJQi4Ls'I{8cTDށD,$U3H 9&ELlЅ̕+b`&W)p{<8mƏuN3}mkio[A\h0v٫4'8W9%xXS!чà | ݙ(9D7tCHk^#O?U~ߣloB0v`Nl]^#in+߿d YR71P7v:κ wCCsDI4%aﵨ']"4b( Ac" }`, ]&4bQ'-ɥxg0;CGۑ`_wO[-@daiW}uS[Et)Y j^n}r`_u\dcI@}3`& H1`h s/ڸ40ESO%ZTU{Z\Oy֧@}H[3w/H:[] ꔾ~@nѮ*Cƈ J Siw +ӠJe*1h%Dz".[gZp y#[/3dQ.W'+^˻ϱN`]=<\^//cY)6^Y{^%[mz_1AIq#Xhx2an!h#xcٕD ͱc){ &Cyc@@?<6}c{4 iS%,.ne}Nt5uyۉI'KQUVEʰѐϺ=R= 'ctrE:n*B 9X{իR(<tw\r<͝x C&vHL(aUH=Q0{C&Yn8,L01mLGyw;`7 02T8ZO:?=2fon #*j䶖\QAld?//6 <;it}08A^>QB E}9^|˲1D1+rE: *w1NaqRĎhν3*0{;G7SoOeU[GL׺;xC4ɼuv74n s M}^-(5 4=֕8:yK.U|5h9aVbaP N nRLhj4Nmp=E7KPJ>As5qKPN`U΁uMn 3MK fP.pqGmB0F\߁X>dAQdNnj-gzѳ,o^RJzw忤g,N~j3jĻ48 ۿqxU_⨥Ʃ4/Ŷ{P[f;ݵ=XN({}dbԠ=S3NS.DD4mT= %r#1qϔ9'2h&M\<'м"Z^\ zT2VѡBzJ?Եb^bafߏ1";>k=:5ϴϾ<%mf@.GAh$B֩X cU wy5(wǧOꡭ.(=4~ Ab½L 1*"Q|#oCИf:]?e-9ū6*\r}JU5s| J`/Uo;I|z׿9ljID?b^ ddYuMYwʏ!iW"CJƨRBOpcߙGdIV5d&pT[{lEw4ULAR$֭ }.u4-Y7G_(dX]՚BR=-߆6Pdn64vmMf eLɃP\t3043Z2%9̎r*6L>aIYi+ Ў'BVϹpyf!{rJ׭jT>\(Ux` ,뇳U+lzwb/5x/I[^^|..b\˫oC O?}(g+.?}ohߟܼW]}ZP _n1hsi'$4I>V؉4PoLo~2U,e{/oaoύb+m ,RhܐN{7On{:UrQ}+kjCd#'e*]R UA[1&*8L8Jfo3cFg%`t=r8~~mzn֌7d#g ݻTϣ"*8S#]moG+Dpk߫K~=bw_ѯbR$N _5EICggh)GGv(rf]]-a"dDXr(Ȓ/2KΔj^5*881o!c i{:(м֪c*Ħ6Gu1FN$yRFqPl0GhV7u=+$TeA^SdN,ɲq;4ѧ7謁T<ݱmeYs7v77V<(eY.W6o9[V59_YtM.Z ],JlyXus>dڋrQ@i "d JVK/1NEthPQe!=OʴH'4O:eݲJo= oD> =$Lse#EwH>)Par,5.2_xYbtZH@/buG096}ǠvOzO:f)X ϟ9wZ~Б?}j[]o*VI B=UR/LTv˻_QtٌHK Z@)fǨi _` is .慪=mj :yP^!+&h՚=I1 4ȷ$ʟTOnb]u.nZNcѡiJfMB}ao$پ(Xg&#HW= p :G<dQ!~(uNGơ9L0^Aʨq~4A(vuyq\ 4<O0nkB,^n7ڌ9*DfT~GzݿSkZ֪ˎ1c eu4fNz}Ģ/S[E"Dž6`ytO ’&nliۺUՔH&d[$\,M^c9R5segW~W/chd)2Bοk0:`YHvGgɴMy=A0Kp _s~v+gf۹ڭڴM©%α[ ᐒœZi-ԑ,lqy 0RNPҹ;fdciaYi"SvmxaXKBɾ(ro:_xdwi Ƶ ɖd}},\(5X}al.E /S'#"WȽh<ؑ[s,#S0c=sxc7)c0dl݀!'RS璉|F@>]==+m<}.[tT6hy%{ hsO-'kWi;9z~̑C'γߊ]]e̿~=yWg^ h0jxwm4}  y [{5-^{e{D=/?Q?iޖbV6GUR%r?@Js˱*L&Z@DM ώ[DbiШ=~Z)cfi.]~Y^\t*s1 0)/Ry"E+syJ!GYѤ:}{ڠ7(b07z[2g}jIKaym6 -¹J :cu3'M `chjrzmUfT'0Vq%=! Œ.eu`Iduޣ";" c> tm][g«HZꜲHyh h$UD}8`HWې}x]+adH yțXa;=-˴/*W]AISy֟'ʁk^[|'`h#Z :td!ݓA H<꾩VbQ[aNZ:F^d)K$Ď/OHh]4HVBp= _^ް_>Wjwt\ݤJ#'[%NC(jZATg+8Fd6-K =9J% 68Zݕ<yb833G eĜ+|þ~x0he+tE7[k>lT\_tNS8)Њ9HB!^7Lm%kƺؕG,H?ŋ5t8OwOp O_ P`>q|ރ/W{bf.qA^'I*fP=^OG3Q 5h穆1 3vm=;mBקJжyRM)BhןkE,/ikey܎S=3V&h&9f߲=0dygfmV[ FH.AB1]Rmu1%+e3祃H`t&Hڇv xGE6a'!Håגi'+cɡ2Ҁ`杌DD{譐<<4FQ)TR*l!$ӟаo"?o=i釹׿lwECLyм3Բn¯)s( IzQz"iAx܁)gyT\AD+Kst3>KzQbjibrRt6+5|kBoU{_7 T'iNiՠ,{xBZ γBևzƆғC7XY}R\ņ( cY JKQ~s:CЪJsF]7KvZ7 ;ph[U[ϼ5vElj]U8e&}&GFqi<[Pؔ y3W@BDǾL}Ց&U9!zETo;cqܵg,>༷LE*1&y#IG㸊1bQ;" hy_VI~*Ƀh1HZg`Bᖙ Ϗf,c璜&CN<7V%zABCvHl<-5 E@7(v>Q21fRm"aR , (ʥ !e(`xE!Ј4avܓ@AuҺ)ng"謙-FD'RsmOsR!+gfńy~?{|#7=Gnp  7Wѽkgu@ ZVj!X&Ws|Z~֠Uqq˛3y?>,WK X-E)9ODEjFѩeRhLٝQx4Cx[ `dv Z^d&Z#]G1ZZCԤX丟<#ߥ)gȊv(',!2ђ>fNy&á9VͽӢk0E5TVWaX$?s,?,ɶ-R/8$5]L-Wf{XWձF&CGנW|AX5Xa Xӫ>?{rY|/-~~Az=悱J܅~/tuJMB` 4h\TILN!槕l[?/sE:Q >ѱY~|KG?ƭnm<\Ӳ=ԊK#sbY!ED-?N,Jo2-~B.HlB? {11~;sD)bcr`㐉^9=!(Lm>a9ԖKV:1ig)NqS,nDZ $US^B&|j vY}LWfm^;t,LǓQS:uz/ɁJ`8*'e#@{U7~ X|:hFj~Gj8d&oF 2hkx҇W9ǬcDh yYp_#Ӱa4 &)aiL[RYK-AQ$))XKYP+Ҹu 52S8% ȼ] NȺ  zuJ>d9'tq.hUH/!-*"Lв~vfzʸS#b6/52(x%Q@9ʻ EN¥ ֖EI0g#f{ݑJz}ё-jp9Qft:l}ù8zLARJ I̤.-sFeo2 j 'bFby*!ÈT88_ ^XT'(#}p(l!f/ 'aE/gpi{{ j&e6Gcb\yBDNQ (d89qQ# y:7f\;YǏz;fWaHRtne[FJah wL7П%Mdi_xѽŪݭEGr*6L^tI6K0F'QH3Z)`/iMǬ0y#廏/l̻5g!H`8cÞڰTç_1.p:2ha(_U>b,,O_t$d"SH*'`ZZ2#+B,fSF=ҡ/5F(el\\]o[rW?e0-Pnދ(4(u-K$;v,+tsM{|D ߴ\Ek-w7~4pS +_ji-xgfZ65jrS1-,Tf݁ϞX6C>)rOI̯U9U-`yl}?/ܽ;4J%]5w\O@M>xP#`\0u|sTjcleKY# F٣J84" ]uя͢|ף?ߢvPKV^*bn[|׋17zg!rWtCkʜnFpa`&Odw,}n}KT4ռg 5 9rmVJc+ƻOd{6} |%rIi,SLKcOC RSެn{hG1)::!ȴ h0▢LxLzo}l_'yV=QՊ< 2AU9n2S4?:0yMsOXƝq z{ |N;$Zǚf =o2F/bqDM{jb-ғ= >vn_x =bgbbNj|'+>w|m l>uܵW}mqxZz`A ܡyvĿ $dB |M(j,M+ओuX>\Z :g9[J$J*K\JFSG2;jՂi'飌No=Cip)% 6PIXQD/5 M Mlb6bVh>4H]'C!C^:='3H(e*<$ew`HMg]B9Hp-C[[;|eA x+0&KP&ir 96]0 Ex)\˓8,$ݮY:ys׊(?vQ*B+lv"r?I1Ave#F6`Aj/YQq?Sq[*]n$+H.]D[rtg<|ufE0YHCNh̝;2Q ҅HIUcb[TŤaNjZIX섵J;b񧰅18n1fh8"]\mauBuozkAQn5Lt3 fk7t(qm|$V}^Lt\q^[M!"W#,W%2,^U<բmF9Uk>zr DSREl*UjjdTCX;;km"$8s-7Kuu scgH#E{w^[0jfg R!-8&bI:vY6+oE֢]"2)vL۝00{T J(KiFUZNmw$!U\\lʬުL*nL XVY:߲o{gSu= A#+mb$~U*-UA 5XFƪU^!DsfQ<{ɵ@k=b{*=|!ᦙ%>KZÖDn|4{na=+rKƟEdsOزd)(\tCNX5M>w ލs|yqfIoF0 & `Q JHS !7{o}Tbu HU ?$i6TԠ3GUelxFlU]q&fN`sz{yЦe\ټb~9iR a dL!Pէ0G廓I5$ _ qRFA.h٥4%m!XQr WT _^Љ$Fb ّR SS:z+=(9d-2CŴ_Z`<䎾`"My5^hxvt|̉8l ')_P*̊ EmeDJj45F!Ɓe#\몳^7F'4cL^Q>>0tX@.="]T{wӁ{(F«>xԈA?܄pk)6jH* TͮD"PBcȣtdU6%Jfʜ`^XGI*g:$0Css+7W a Ͻ}[_}|= ClygЎ0-'$h6d &\VdSy"QUAՠDeesBNr/m.$i>i]`>'#.zw^e'Lߗx>]늘P^^,7C) UϺLA5GΙ֝P&pv +`/[֢Shz~aTQ(fgӸFYFivi}:O?R O')?ۿ丰kt_r,C#-Y@Hg>/NjE~Ku69ΊOXW]>l:Bpq)X-4l-lF*[@Zm!G*᦯|Y+V(_KIXlr}YFw~4{{opMy:w/1wٲA0l7_wޖ.q/4o?ykc~/tgd?|.U>W KH1H>%jP>k#^{iYٔntb//U5'k]C/ T@dwZqƓ>|PY)N:5kR\My8.2Zq>]C`~VeՎQOJja\UhEF?cz1rY$0ΖltTG?oN G*z=ZvګotB}y7ت J&VWPA;v*`Њjt){|bqh:z3Fw7t=|/Fo5Jہ7GB4Fwd|XfE%_[Feߟ 7GHq*WIoX[:~,1Yb 5gyc~VsF@mrqz$/Hqg"*D|sG xo㔔~,n ;"+<zݍˇXv#wI;VkxLk{ssd$!d6VĪxrE0-t_'HXX+EbuP]< KMRYIV͂9=q64B~~?_~rv^ W&W讏Rj}_y6Z\sI'GGүY.7S0K@I&Aytd>@`k5I.u`ۦj-9iW*|v8xƃIJ ,%R߮M}{/-ۏC& :*i~o-֖IiXWF{!\0^g?E@k-"^Ipz}!tW6ڇ_Le92i-y`oN#h E"fKMZ.c̉Aֵhgܩ>7,XB%0fwc>Hf~KH7{Ճ;5Es(vw|u8 狴<W?hp8ft=:;W ~Z ç;C[4#(ww Pьbœl;mOngPhiG~G5[%n5Bj}3TF,vleZ7}Upoe d7 znTg5bg7?ns+ou WrWqZ݃-uzyo^&'&Lc쉉?<fv>3j2ttFnV]6X DTͪ~P(BML)ڜRR/-6j6V|ղCrAz1+!۠RÉ.a{14؊"R2N6*2rV>c+@qd\kHK0h!m<"0nV#5d(NeUqxX [E\b U\3˼zel$< O ]/q"D0bhYlp1䅶N;q<@ۣhE9dvpP[r۵=b93ޓ@՝GzݒĆ|7 -$VY7-aiL Dmhq}szD69b {1%*7e] cpJE>QadqT̠m1 ʲ8MeKdRh$ u84#+>v5ⴤ ՠeT)J1XfȮ%k D#cA'B1-ڮVh G;ƐC!wLg3:͇S{>/D4̊?~vԫ=.8@`iݚpP٠q-UlŤcѲK$0h ΖΩ$&,(߈z`&CN}AE2mL7?ī>ViūEN]S:,[vC|v -ԒPY|ؒ18莐OKb#zsח *h>&ks-&aF>5- \ݼaLpJLY~n$uO}'NLt&>}LG>ERzIh m*ICCݓ%-ĄVO D3+m~_m_Y1AX Xwy!BV>u8m]W_Z%$ vvР6Y :d9RI^y^d~~hᐦgNFC`4G8Ҿ]]~9Ϡno˭,awZD_XY '4h11-r6Q9wP-զ"t6[Tu.VQ LtCHd)_GBr󒏾x!d_bT^EQ^H6Fܠ?cl iʡ %F̾t K,؜C0t:E;TfQl;Hr4icGpFXLڣCQ$PSPajy"|xBuh$/Hb&v&?;2xPA5h9v7Qt9fѺ0b EZ:g?`yX R-pvXMm|[KlE*P1Y=EchUT)[ Xˀ5+x +O.6Ϣ4X'|*wVdȼ$cs{–Љ $ BF,.M!SzF:ޕ Rv'#h;*__.r:c ])QSF1{B9҆N6Kڌ`I}oA"݆jus9O_uPjhw?^LBAgF ɤb ^iZr]<☌)1f5YԨE DKۓۮ_ E»,a>\L_Z6Oˊ dJXE ŧeG|mMyD$2g7o E>7E1uBK2Y/:f|-n98vHnܾ=r!}OI8ocTg>FVnNgˡEY AU_.KY p8\߿7m?;%k?99d3(doZR2IDFq$Bt1{!8̶VC('A!Q DbX<%( O E#ѣ6p3=^/j8e|Lþ7N¤=9ő2/by KwjUD3>K0}A!h$vh→_Hm]յ=ZHY}_NnxG,0v@};|77 a6s@?g:eZڱyN1N;^qOc8r_4kݛ}87pq0'9qǐicO.1]n7 ;5Sz]}-rQ\u;xF*OPEo]\~ [SY#ͬ?}{2;)$s)s"ju-A_S^[l]1B/PN}ukz_]F ?pwĕ+^k߂Cj/[u}AL-?#pQyG/`gٙ50c$(odGuR6Ux/2#"32"ep"Ubߒ{DRޮwIm:? A\k>קDh)ba8~Rq.-Q {]+ o/ދUCO9e0 qet\OZ9{y cnuLQ11%[*\kTWտ ږS#J# &y yJY%gLJ` Jmˀx͙D)9 ߮ʍR+?-M٨xt?#& 76qL6kgD>Дz垚K!@kڛD4jP{/6Aś>_dR>w/k !GDPj0=Bj amPWHؠo;-Hp?`y%+; W^tU 4'ה&i `vZ3av\V6kI8KTdy`ZZNrJwRPDd_*we3.A YBUr2!9K=(CrR;:2To>֖oQ{&TjBf /#MTh?FƧ+9gƇ~?w ]S!/5Yd˪x^ g?|:j`FAè.+k8I@Uh}}{ag^=N SyC{)pt1OnN+:xDXl"a#254A ϡ<崰F[16F$*\ a vSITL}hy3 omxR8Ki FS>L'/pّh-#:fcK3aꈦQȞ?[x+y0 ^'I3' ]s/qkh-ʹEĦzyɕC*mb&4ɳ_󠭑1ϙ+(TP0 *(4K:](=svA!f* 1KcCoڷ+!7ɵFެ?R鏢Kr=#qJ~)A6bk:| M??M{koU[5_ϖ[2OtOH|d:ź@@+)7=̭mւT"Ug o3ع?dKxG tD?=fSq߿l$3Z8_?w &J‡n%LyQdǻ]䒔 ǵE'ǛtA"e) :Lc~ctfK>*C/ u|PqAĹVi!qٴZ6|lZ g g8hR{\̞t ӣ; (q3 L ۋsuqpIZfb2?WbDul^H>qW8ZS8:"Q!+:`½Km2Və"gܱ zH׬!C+@r3nKluLjՎ`+]Iki^1hYRo|dRKLѫEЁ/J2h0  2-5B’2*I!=r<m5PVZ_b`Ғ?ez4o pFQKä]BF(8;t|$F'T2s/ݢp\ղWB)l*8;礳 .Al/g|})id;߻LFo5 Bj)8tΥϣR28 29e Mњ]K %[bxQr\$js:N)vr_gCN)| Qg5Ϲ`|.'h$jqŦ@#.:B@%ڄLj:[cj.p͚B*.="D͖"5Ŗ@.4bE3"\+Ɣ^Y V !N-j8Pz.y9װAE+NTl'=zxl~:u0-toحEH 5R1V[|qD^xBH*%מ{oU;c*aUT:cOO%vЬ2vf X20w4>R+nG]v cNJh Q;N4,.%k492OO7S-B3Mō4"lUL4JPu3_*/v&upF~DD*@ia]" d (´16뼌hM9GE$:sE<@NI@Hz9=ʡ :ϊ)mKg8ߏYqqw1?x;yO3lgf9"0 \s[$׋{ (骑8ǽ FN ϼ#y sY(  Q6-f[jAwNzW0a'%pﴄPN"j .,R8Y- ӨK1"az`F]]S! Rj5q!K(/gޓ\ 1$ \dF=v'ݞhvR#Eb7e COFu ɓ]2*h[$PУi"ٸ>zyvYYO08GR%wTw?OOfB.{dTmX"z:o@=M'xTo]\EC~cȷf&I|/ .ZJ^6@ o#yRjSZ\{M!gl]2 [!EpTW^2fku3,O\4Z]@zv_@k†[]8kѧ6oyRO_K@7[ )`zo! bOaЭĶߌng:G7?;˟Tۛ;{'dbhv +;qTeT)g.I96dOա#,l\)\UY91^W̨RjCf0A%1ણA_>3I*e0a҉*aYfӛiʗǰ?e{_;:CoF *5?i@uW=8%Z[VGpt_w%y7hB΀ {63srPxq&KAp ]i=^Pͨ)]߫v+lkfVg^oSTsB=8iCHdMHHp?9p)Fk}x^dحuuWqrdrMƭj[(\n'jl[-;O&$&{=iEh3I3 r;K6?VƁ\aT-`-ag2P^f.QLZw\DjϴՑ,"@ GȹQGUJ."TH= RSZikAMd.Nxnh ǍtT'yiƥ.<0N'GEL~yx$HQ\K7pHFjJoC)É i;p k"Y“2ZoQM"L\pOs)BT‡ 9cm2&:l,~Vycc }.uƵB)>uqJOnV-o~e w%zU+تĢ6}dBܴR..q>pt!knlO}Nf_d`8Z cKqЪNMKZBn_Ղc 2RvcMASI%1DBhFT17\B}Iq?c ;Ц,vvw|dC=\eѿ*yJ5Rj#lmah}g\:+CQ?^'p*6ͦX\vE0 Uw!˂d]LWsX6Gr"1I~x΁"v(wFchb{TPųzt9zLis[NA?a wwO\7`2lLrcSLe颥ʻ<\[Rqp^ݦxꣳ_ˀiFR+Eʂ=8PȲP) %"my .3rr [eߠr`.2>}`ΘUv&~ ?ڰ(a N]N)_Q֬gW .UZ.7P(0%N-^#@u2 OOe. TH8 e j orIZMoGŤOrȮgQτյNtD7bRBV0q&]gNFw9 @EYoZA n:xӨ5@"[!"B$> 2XC25 KPJ\+IyTS2V&$qt+K $ SYQ؀5c+Q+ˣȤaj\%C^`9|p+S+w>{_0ZzLYF"h%K:?4< F[(Œ;^20zX0@;3\yϯoF)Ө@1pJ[ݞ@Jt=ܣK6AF](| 'Q;Ո"Ҷ>ZvT-m72ⴙ.u͸O^Z ^~8Ǚ v Z¿&x,-nRn-H9#Fg135("4њ6b7oDZx10vc)g,] 9T "H[)`)k#AԂB RN3~ڻ-r=?a dzì eDgֵ=9, D2[rJGx_w_톉D(,#.G}oR??ã>çƓlk4wW36KeUGG_ thlb8 9єT2/~y/:)_UomSsv<UNi T*c3 bƍu&'Yg #2YD雩$+FJ|(-l) xiҵ ?u~tZimVoVoVoD0ZHf2!8;T iSxkoVnVns57he\κ7/ qa=0bUޛW.Կs郙E$+e$"(mاoƧЪӊJ4k9Xd)|NdžD)`q*}ykQJA}~]`#a#=Flc!7۟+L6!Df)!DSKiZk_ѦPE,6x?/>}9THyF:Nu$|d 7IÃXPggA(܎;GD$  lhh{ViEKR=ZWjɝdGQ2e%D$cU"1 #T34ZNO ,?./N/,dS8YN9eRJsYz]Q0G >ׂy*+~.=2,va05۸Kagy̯KCr V>ۇIQ ]sZbe)R g){Xz`n|6nǦ1rMuriBhxRɇI-\Zc<oIZ"iEzW4I-(!4Lɉ/Ǔp{rxe=H+]ŀT 5/@EJ$d#p aɂO֝lKo]0Ck%Z^ި<̱ƪ @B Pb_+DUk{ͬբO@07d@Z'DW7ql RÍeY$CX2Dh`<C JÕUK!8 mUJ".@DaUdu\дIDDضm[\ZQx3/wvP{Ѯ) kwZ]`wg2ֺha#)Q;yD7sDKX<ѥZLMkifE%B(Ш(L^Yli?ٟgū5}\Uw-FkԂ\q<-wN`%̹v0Uk:'oH^jìH.IӚ C)2 O@|F>ςfg̐-nMq-O;P }u>:_9rSqq#5bd7ͤNAA 7rېfr QN냰q* ^UĤV6)0Ho4HE+4%)F*yioD0* D |鄡,WYef"i"`L$7}`V}Ԋkx1B(fBT`d1)T!iŔSQf~cnΨ)kPy?zQVU 堭#4k yc E;J+4ve`4Tp:\RxNBx3NBP j M4ݽӂўMhde%HZǛz?? GgߓԜ=@gYӌ5EZ؝8yQ$X?=Ts}s㌿菇:8.EZ>HR4UT<>g/׻A~!{kGп= kHQ`f5")2EyW/Q؍D0cz$U$ ecguIb/c)#<3Hy9mO @KN.X_n>g aØXLd˲њϻ0֣].YUnHq}w' <|6wkfENHL׻&_er@u+/܉]  J/)  F f(`u(0j&W_| SDZOa<n8VNN8ոj\iJtnm[$CѠ z ST4zRG.f)JE-ƀh2J4E8Xw[ B9!TSuDzIVTkPygAi{ "Y[UEh7N#5]Ŋy>$yUS3&&%IP"ȅ=qwXwX[#szE:M@a!}@~BR>Cs -vMh#ƹo#aF- +kg?eMeRYs,;' 6M~/<ϾK)J8hJ)Ju~P坽 Y,G DhZ@]L|M)A `|6`fQe'_k 6+ߕq VIς33(0.`k[wk_ ЅLu2Z>o;E"oG.㚧Vf|\^g&Tx`%)p;x+458W:ڵ%;vML{n݀K4M鬣ےiM[ ђ OlJwgNJt B0myXx,@'"[UuyXB]4JF6y#5 p6;fy(TnjWu#o~(l kWU5:Zr߱pm|uU)w&͠p8ߪr ;+8^ofXNw(rO 4/{TwF[TR!_Sއg0e[.)C']RAѬ[vGS[ќ1~=F: 떋A}Gv:umAh-֭‡h1oYFEHpl?aNC",ϫM !0 6ΖޑpTOL @HYM *"6&|MRun=PY|T3 Isu?& GrĈË+ɒ<":Nx^D[(''TJLT$a;Ts*T>b,lwsDh6a0îhE'GƂKTm&΁!.VUMҫm䞛UuŬW {^_B;b[\{K@96 XakH(-VZD*1Edž+ƽ%%jX9Je>iCc$PEɨ&m6 8 lٖLɄʹhDeNP`;RɉaO AJ I]hyS!OPK)@ա'aJϑn2͎ӭS~ݢ<5ߣR/F[uOS*k+Ի*XrU %~UǸlt|šݛSB[{sGuhuKU~*ҿ ݻ6w擙z0rbӁkhaԐѳI)||ϊAz2^/k~;qfs*嶎,jwH|5惛>>Wгdo:۰`艙/Ivc/t, a@ zآAq9#O }G rd8gOp?H|ͯ+zR&!X Ϡ8sV`_a:{&`H)okHMK0Ec տqv n2!W!cxe!5(A+ Q~;5Mn2uU9ޟj[mZ&γ? 'p)ۊ஦zzCz5G,%`>x, (Z0VuԄ"Ǟ363ӌ*fy$>ӂҖaa&Hp,^p^Oؑ8P iz; )`6ݹ> K"=1 rI R@8dU80!! 1Z9l-+ qK)#R@bSָ55KO4QņQ͍ #݀[YE swUeH5੖Y%'g9m103J56|nhˍWF9z*`85Ԃ&Hm_O9tƚZ_4}uJrSIkQ|;;U=NRF=_-\ Džl_wBybyu]J]z%x"O( {V2Y8![&J0^#:^E^ŠеZ.QwK30LN6Q)U c~t}~E= ]\: FShd'nc.A@[G ߳ɂU `d}ޅF"y5g$ATWT U˃UT̃L B)V*éB+YKgrb뵪 IFHa[j3'&8wB'u8:J'^YN Ӻշ߄C SJO~jw嫆snGU,f/fH\McVR2CK'uI^&#)16XFZ$S^zNU5Р53kb pN"%jDkV[i(g rt̿A>,٠FIzlOMuQSĆrs07 3ʭԅQi@{% HtA`]yE]$ߘOb ۇu.&hiޭ܂Em\xT֛ߑFf_,G^] ,ADT(h ]˵~kH߽sз͟oz%JɟnnB(!قb wcɆJC/o?&?*lZöfwr+<"X 5o{= B:>tz}[`"V+/L͝fI_O/|RoR~ߺJHkΌJ~^a~J}:Hܾ 1=opƭ5-jAJz9cۦ]QeqMM;ra'E.?c%t_s k] -XDM2iNu Dl^]4D'mR̴0̏jkbtmm^}m1,\kzt-Pq.嬝$d X*UvǔhQ 1/5ʖ[n ʯ 4`P Q:rPp J42X9Ē+gew} ^M#y2$w3rsAv8yI6 gt'XψAc|pp3f3{K 'uvXo y"eqAv. gGH RCEsr3 :MܚmqVx0;1x[dRj<,F)ReF wT5Vs0)oS8ß1tO~]:&F?݃' / ^M-6ŮpGEjFx=M,֥LJy$ *W~<;!DPt5=o!2BA@0y:ۋEK *j{h#+_ 5P2 ߋ{S$9.w)pbzg)Tx5i%H1zQiXó5JQS?Ly'&:M0q3GqoiW23qƋb3>oz+'[V UO@U &QܻkwY*! ?}q^,\7%G}_ET.U?nMER~?^f'pt>*zatNN6qd|6+O:]fw1ow JwՆ~J0^:龞2$+N"|;;GRB(q Yկj^׶T>PWtfˋĺq\N9O?mY3fr̲ǷG;zb G|_ӌddUcWKdY+_.>$5OڙL^,T9,iN,tĞZ@:j&WD-kRNo>ZcN*59bMbJXL]o JT RkgbD/7Ȩ6[N@';>4jyCc9-ЫJ^-2:ppCAu@!{F,e7!!jh$XN{Ҋ _ k0`7R).srOc=cPf rƀl4yn6#⻻ē|/YTAjqhҠAP"u6RDA;T#aT!BH+`YLEh;G_dRS\21jSy$ќvZhVT 5̡8tFHxp(AQBJ)yj 4p^2] dCmsϩ}v4=Q*xm0]O0 K*3spdn4-|gI[%llr׋e싻\Ȝ9Z`I+t]D_@UBegj?]WBJAuhtIzPk7d1eXf/1isfrM ?u^yw|]w#ۮl]a?ON{)ƫxU*U52BcK WP u x7:+$ .W&PQΐH#O96O+&w0iQGӏvDu]43ѓ_WhB@w>ЕGb>u 5zh Y1zvyں<-A6!.O.Ok^}ȑr@×jjuGc*Wo HuZj2C pSoѪQ:y"&* \QFSOk)Dpڥ74H-K@c鋿z~٨ι[L*^7з8_giTx׸g.G Khv`MQp.[HGy<9^$dP  L^?n-Cކ=^Ղ; .f${n$Q%01$5FEƺ`2LܶXT!1Z/a˂3꼦VP@!g% ׎XKG)I\r,5_^ܝUh c] ͙try%y?6ZR+7g0RoזhQo;%#6)%y>P$3D1xxލ',>CT"`UDŕJf+,*?y_i+4%4DG3 ;pAcftX-\j-@GW6`6T '۳&x+ٓEE1N_(6ߧF i]j G ]MvXj/GMC(bhQT)D̂ ueZD&-Z#8Rn+dw\t/\[ ɖE`V0$xq<>\?0eZbeڥ? BA8w?[n?51D0nz\f\~DCw۫ޡeq;˥Cd;G%'MWГ6C[yiV 2ƿK1GmڙQ/΀Tw\N;wPP%+x`VT20 ./LC5zXPEEb&OLྴDN=%%(Ӗ.4 o avar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004050064715154764155017720 0ustar rootrootMar 13 09:12:38 crc systemd[1]: Starting Kubernetes Kubelet... Mar 13 09:12:38 crc restorecon[4681]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:38 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 09:12:39 crc restorecon[4681]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 13 09:12:39 crc restorecon[4681]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 13 09:12:39 crc kubenswrapper[4930]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 09:12:39 crc kubenswrapper[4930]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 13 09:12:39 crc kubenswrapper[4930]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 09:12:39 crc kubenswrapper[4930]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 09:12:39 crc kubenswrapper[4930]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 13 09:12:39 crc kubenswrapper[4930]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.706287 4930 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712190 4930 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712219 4930 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712229 4930 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712239 4930 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712248 4930 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712256 4930 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712264 4930 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712275 4930 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712297 4930 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712306 4930 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712314 4930 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712322 4930 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712329 4930 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712338 4930 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712346 4930 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712353 4930 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712361 4930 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712369 4930 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712376 4930 feature_gate.go:330] unrecognized feature gate: Example Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712383 4930 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712393 4930 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712401 4930 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712409 4930 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712418 4930 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712425 4930 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712457 4930 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712465 4930 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712473 4930 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712480 4930 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712491 4930 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712502 4930 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712656 4930 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712666 4930 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712676 4930 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712684 4930 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712693 4930 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712700 4930 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712708 4930 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712716 4930 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712723 4930 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712731 4930 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712739 4930 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712746 4930 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712754 4930 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712762 4930 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712769 4930 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712777 4930 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712785 4930 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712792 4930 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712800 4930 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712807 4930 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712815 4930 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712822 4930 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712830 4930 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712837 4930 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712859 4930 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712873 4930 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712883 4930 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712892 4930 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712900 4930 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712908 4930 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712916 4930 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712924 4930 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712934 4930 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712942 4930 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712949 4930 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712959 4930 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712969 4930 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712978 4930 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712987 4930 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.712995 4930 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715115 4930 flags.go:64] FLAG: --address="0.0.0.0" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715141 4930 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715161 4930 flags.go:64] FLAG: --anonymous-auth="true" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715172 4930 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715183 4930 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715193 4930 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715205 4930 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715216 4930 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715225 4930 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715234 4930 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715244 4930 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715254 4930 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715263 4930 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715272 4930 flags.go:64] FLAG: --cgroup-root="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715281 4930 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715290 4930 flags.go:64] FLAG: --client-ca-file="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715299 4930 flags.go:64] FLAG: --cloud-config="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715308 4930 flags.go:64] FLAG: --cloud-provider="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715316 4930 flags.go:64] FLAG: --cluster-dns="[]" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715326 4930 flags.go:64] FLAG: --cluster-domain="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715335 4930 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715344 4930 flags.go:64] FLAG: --config-dir="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715353 4930 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715363 4930 flags.go:64] FLAG: --container-log-max-files="5" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715373 4930 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715383 4930 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715392 4930 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715403 4930 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715412 4930 flags.go:64] FLAG: --contention-profiling="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715537 4930 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715552 4930 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715565 4930 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715576 4930 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715588 4930 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715598 4930 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715607 4930 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715616 4930 flags.go:64] FLAG: --enable-load-reader="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715625 4930 flags.go:64] FLAG: --enable-server="true" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715634 4930 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715646 4930 flags.go:64] FLAG: --event-burst="100" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715655 4930 flags.go:64] FLAG: --event-qps="50" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715663 4930 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715673 4930 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715682 4930 flags.go:64] FLAG: --eviction-hard="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715693 4930 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715702 4930 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715710 4930 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715719 4930 flags.go:64] FLAG: --eviction-soft="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715728 4930 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715737 4930 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715746 4930 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715754 4930 flags.go:64] FLAG: --experimental-mounter-path="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715763 4930 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715772 4930 flags.go:64] FLAG: --fail-swap-on="true" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715781 4930 flags.go:64] FLAG: --feature-gates="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715792 4930 flags.go:64] FLAG: --file-check-frequency="20s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715801 4930 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715810 4930 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715820 4930 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715829 4930 flags.go:64] FLAG: --healthz-port="10248" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715838 4930 flags.go:64] FLAG: --help="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715847 4930 flags.go:64] FLAG: --hostname-override="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715855 4930 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715864 4930 flags.go:64] FLAG: --http-check-frequency="20s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715874 4930 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715883 4930 flags.go:64] FLAG: --image-credential-provider-config="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715891 4930 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715900 4930 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715908 4930 flags.go:64] FLAG: --image-service-endpoint="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715917 4930 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715927 4930 flags.go:64] FLAG: --kube-api-burst="100" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715936 4930 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715945 4930 flags.go:64] FLAG: --kube-api-qps="50" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715954 4930 flags.go:64] FLAG: --kube-reserved="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715963 4930 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715972 4930 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715981 4930 flags.go:64] FLAG: --kubelet-cgroups="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715990 4930 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.715999 4930 flags.go:64] FLAG: --lock-file="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716007 4930 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716016 4930 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716025 4930 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716038 4930 flags.go:64] FLAG: --log-json-split-stream="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716047 4930 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716056 4930 flags.go:64] FLAG: --log-text-split-stream="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716065 4930 flags.go:64] FLAG: --logging-format="text" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716073 4930 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716083 4930 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716091 4930 flags.go:64] FLAG: --manifest-url="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716101 4930 flags.go:64] FLAG: --manifest-url-header="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716112 4930 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716121 4930 flags.go:64] FLAG: --max-open-files="1000000" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716132 4930 flags.go:64] FLAG: --max-pods="110" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716141 4930 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716150 4930 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716160 4930 flags.go:64] FLAG: --memory-manager-policy="None" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716168 4930 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716177 4930 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716186 4930 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716195 4930 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716215 4930 flags.go:64] FLAG: --node-status-max-images="50" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716224 4930 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716233 4930 flags.go:64] FLAG: --oom-score-adj="-999" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716241 4930 flags.go:64] FLAG: --pod-cidr="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716250 4930 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716263 4930 flags.go:64] FLAG: --pod-manifest-path="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716274 4930 flags.go:64] FLAG: --pod-max-pids="-1" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716283 4930 flags.go:64] FLAG: --pods-per-core="0" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716292 4930 flags.go:64] FLAG: --port="10250" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716301 4930 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716310 4930 flags.go:64] FLAG: --provider-id="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716318 4930 flags.go:64] FLAG: --qos-reserved="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716327 4930 flags.go:64] FLAG: --read-only-port="10255" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716336 4930 flags.go:64] FLAG: --register-node="true" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716345 4930 flags.go:64] FLAG: --register-schedulable="true" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716355 4930 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716383 4930 flags.go:64] FLAG: --registry-burst="10" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716392 4930 flags.go:64] FLAG: --registry-qps="5" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716400 4930 flags.go:64] FLAG: --reserved-cpus="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716409 4930 flags.go:64] FLAG: --reserved-memory="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716420 4930 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716429 4930 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716464 4930 flags.go:64] FLAG: --rotate-certificates="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716473 4930 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716656 4930 flags.go:64] FLAG: --runonce="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716665 4930 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716675 4930 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716684 4930 flags.go:64] FLAG: --seccomp-default="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716692 4930 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716701 4930 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716710 4930 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716720 4930 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716729 4930 flags.go:64] FLAG: --storage-driver-password="root" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716738 4930 flags.go:64] FLAG: --storage-driver-secure="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716746 4930 flags.go:64] FLAG: --storage-driver-table="stats" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716755 4930 flags.go:64] FLAG: --storage-driver-user="root" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716764 4930 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716773 4930 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716783 4930 flags.go:64] FLAG: --system-cgroups="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716791 4930 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716805 4930 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716814 4930 flags.go:64] FLAG: --tls-cert-file="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716823 4930 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716834 4930 flags.go:64] FLAG: --tls-min-version="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716843 4930 flags.go:64] FLAG: --tls-private-key-file="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716851 4930 flags.go:64] FLAG: --topology-manager-policy="none" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716860 4930 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716869 4930 flags.go:64] FLAG: --topology-manager-scope="container" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716878 4930 flags.go:64] FLAG: --v="2" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716890 4930 flags.go:64] FLAG: --version="false" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716900 4930 flags.go:64] FLAG: --vmodule="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716916 4930 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.716927 4930 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717146 4930 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717159 4930 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717168 4930 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717176 4930 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717185 4930 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717193 4930 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717201 4930 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717208 4930 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717216 4930 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717224 4930 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717232 4930 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717239 4930 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717247 4930 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717254 4930 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717262 4930 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717270 4930 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717278 4930 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717286 4930 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717293 4930 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717301 4930 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717309 4930 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717319 4930 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717329 4930 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717340 4930 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717350 4930 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717361 4930 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717369 4930 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717377 4930 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717385 4930 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717393 4930 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717401 4930 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717409 4930 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717416 4930 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717424 4930 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717454 4930 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717463 4930 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717470 4930 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717478 4930 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717486 4930 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717496 4930 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717505 4930 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717514 4930 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717522 4930 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717530 4930 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717539 4930 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717549 4930 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717557 4930 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717566 4930 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717574 4930 feature_gate.go:330] unrecognized feature gate: Example Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717583 4930 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717591 4930 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717600 4930 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717608 4930 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717615 4930 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717623 4930 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717630 4930 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717638 4930 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717646 4930 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717654 4930 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717662 4930 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717670 4930 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717684 4930 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717692 4930 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717699 4930 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717707 4930 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717717 4930 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717726 4930 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717735 4930 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717743 4930 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717751 4930 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.717759 4930 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.717782 4930 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.730456 4930 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.730509 4930 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730656 4930 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730676 4930 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730687 4930 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730698 4930 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730710 4930 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730724 4930 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730734 4930 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730744 4930 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730753 4930 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730762 4930 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730772 4930 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730781 4930 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730790 4930 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730800 4930 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730809 4930 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730818 4930 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730827 4930 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730835 4930 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730845 4930 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730853 4930 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730862 4930 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730870 4930 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730879 4930 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730887 4930 feature_gate.go:330] unrecognized feature gate: Example Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730897 4930 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730905 4930 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730914 4930 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730926 4930 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730937 4930 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730946 4930 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730956 4930 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730965 4930 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730975 4930 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730985 4930 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.730997 4930 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731007 4930 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731017 4930 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731026 4930 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731034 4930 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731043 4930 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731073 4930 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731081 4930 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731090 4930 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731099 4930 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731107 4930 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731116 4930 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731125 4930 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731138 4930 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731146 4930 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731155 4930 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731163 4930 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731172 4930 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731180 4930 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731189 4930 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731197 4930 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731205 4930 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731213 4930 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731221 4930 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731230 4930 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731239 4930 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731247 4930 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731256 4930 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731264 4930 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731273 4930 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731282 4930 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731291 4930 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731299 4930 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731308 4930 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731316 4930 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731325 4930 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731336 4930 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.731351 4930 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731697 4930 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731716 4930 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731726 4930 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731736 4930 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731745 4930 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731753 4930 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731761 4930 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731770 4930 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731779 4930 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731791 4930 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731803 4930 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731812 4930 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731821 4930 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731830 4930 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731838 4930 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731847 4930 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731855 4930 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731863 4930 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731875 4930 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731886 4930 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731895 4930 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731905 4930 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731914 4930 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731923 4930 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731932 4930 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731940 4930 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731951 4930 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731962 4930 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731972 4930 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731983 4930 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.731993 4930 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732002 4930 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732010 4930 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732019 4930 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732029 4930 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732039 4930 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732048 4930 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732057 4930 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732065 4930 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732074 4930 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732082 4930 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732091 4930 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732099 4930 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732108 4930 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732116 4930 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732125 4930 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732133 4930 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732142 4930 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732153 4930 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732164 4930 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732175 4930 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732186 4930 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732197 4930 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732209 4930 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732221 4930 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732233 4930 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732242 4930 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732251 4930 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732259 4930 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732268 4930 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732276 4930 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732285 4930 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732293 4930 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732302 4930 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732310 4930 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732319 4930 feature_gate.go:330] unrecognized feature gate: Example Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732327 4930 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732336 4930 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732344 4930 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732352 4930 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.732361 4930 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.732375 4930 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.733580 4930 server.go:940] "Client rotation is on, will bootstrap in background" Mar 13 09:12:39 crc kubenswrapper[4930]: E0313 09:12:39.738713 4930 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.743377 4930 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.743838 4930 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.745902 4930 server.go:997] "Starting client certificate rotation" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.745951 4930 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.746126 4930 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.776491 4930 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 09:12:39 crc kubenswrapper[4930]: E0313 09:12:39.780235 4930 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.781148 4930 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.803059 4930 log.go:25] "Validated CRI v1 runtime API" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.848980 4930 log.go:25] "Validated CRI v1 image API" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.851038 4930 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.858214 4930 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-13-09-08-25-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.858265 4930 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.884668 4930 manager.go:217] Machine: {Timestamp:2026-03-13 09:12:39.881872525 +0000 UTC m=+0.631787242 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:f35d26c2-9d58-43d9-b3d1-f4e2250d2ace BootID:577d718c-b49d-4fb4-b481-cae7fc188388 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:4d:ff:95 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:4d:ff:95 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:a2:cd:72 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:9c:1c:9e Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:f1:f3:3d Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:5a:1b:12 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:86:bd:0f:a4:ad:f2 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:fa:09:c6:a4:3e:07 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.884944 4930 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.885101 4930 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.886635 4930 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.886835 4930 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.886872 4930 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.887098 4930 topology_manager.go:138] "Creating topology manager with none policy" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.887109 4930 container_manager_linux.go:303] "Creating device plugin manager" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.887673 4930 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.887707 4930 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.888791 4930 state_mem.go:36] "Initialized new in-memory state store" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.888902 4930 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.893977 4930 kubelet.go:418] "Attempting to sync node with API server" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.894001 4930 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.894029 4930 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.894044 4930 kubelet.go:324] "Adding apiserver pod source" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.894058 4930 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.898045 4930 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.899064 4930 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.900871 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:39 crc kubenswrapper[4930]: E0313 09:12:39.900945 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.900867 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:39 crc kubenswrapper[4930]: E0313 09:12:39.900986 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.903761 4930 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.905822 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.905854 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.905863 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.905872 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.905890 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.905902 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.905913 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.905932 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.905945 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.905954 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.905966 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.905975 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.906901 4930 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.907401 4930 server.go:1280] "Started kubelet" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.908529 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.908554 4930 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.908764 4930 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 13 09:12:39 crc systemd[1]: Started Kubernetes Kubelet. Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.909726 4930 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.910976 4930 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.911041 4930 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.912832 4930 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 13 09:12:39 crc kubenswrapper[4930]: E0313 09:12:39.912856 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.912868 4930 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.912884 4930 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.913586 4930 server.go:460] "Adding debug handlers to kubelet server" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.914088 4930 factory.go:55] Registering systemd factory Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.914140 4930 factory.go:221] Registration of the systemd container factory successfully Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.914303 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:39 crc kubenswrapper[4930]: E0313 09:12:39.914489 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.914693 4930 factory.go:153] Registering CRI-O factory Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.914720 4930 factory.go:221] Registration of the crio container factory successfully Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.914793 4930 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 13 09:12:39 crc kubenswrapper[4930]: E0313 09:12:39.914785 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="200ms" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.914820 4930 factory.go:103] Registering Raw factory Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.914839 4930 manager.go:1196] Started watching for new ooms in manager Mar 13 09:12:39 crc kubenswrapper[4930]: E0313 09:12:39.915188 4930 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.238:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189c5baedf62bc90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.907359888 +0000 UTC m=+0.657274595,LastTimestamp:2026-03-13 09:12:39.907359888 +0000 UTC m=+0.657274595,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.920837 4930 manager.go:319] Starting recovery of all containers Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.929994 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930057 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930076 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930092 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930110 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930127 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930144 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930161 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930178 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930194 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930209 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930224 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930240 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930259 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930278 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930295 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930311 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930326 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930341 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930360 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930412 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930520 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930543 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930559 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930576 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930592 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930613 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930631 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930648 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930664 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930718 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930735 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930751 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930769 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930785 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930802 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930819 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930835 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930853 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930869 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930885 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930902 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930919 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930936 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930951 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930969 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.930986 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931002 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931017 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931033 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931048 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931064 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931087 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931106 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931125 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931142 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931158 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931174 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931190 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931206 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931221 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931236 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931254 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931269 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931283 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931299 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931315 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931331 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931346 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931361 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931378 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931401 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931417 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931470 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931493 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931510 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931526 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931541 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931559 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931575 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931614 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931631 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931646 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931667 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931681 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931698 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931715 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931731 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931747 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931765 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931779 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931795 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931810 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931825 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931877 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931896 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931912 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931927 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931942 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931960 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931976 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.931992 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932009 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932030 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932053 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932073 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932091 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932108 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932126 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932143 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932176 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932194 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932211 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932228 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932250 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932267 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932283 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932298 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932314 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932331 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932347 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932364 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932379 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932394 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932409 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932424 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932463 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932480 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932495 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932510 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932525 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932540 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932556 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932572 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932589 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932608 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932625 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932642 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932658 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932674 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932690 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932707 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932725 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932740 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932756 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932772 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932788 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932804 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932820 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932837 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932853 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932868 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932885 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932900 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932917 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932935 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932950 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932965 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932982 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.932998 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933013 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933028 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933043 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933059 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933075 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933090 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933106 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933125 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933143 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933158 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933176 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933191 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933208 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933223 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933241 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933259 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933275 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933291 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933307 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933323 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933338 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933355 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933371 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933387 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933405 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933421 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933456 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933473 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933489 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933504 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933522 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933537 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933553 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.933568 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935674 4930 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935735 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935762 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935782 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935805 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935827 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935847 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935868 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935888 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935907 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935925 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935944 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935965 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.935985 4930 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.936003 4930 reconstruct.go:97] "Volume reconstruction finished" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.936019 4930 reconciler.go:26] "Reconciler: start to sync state" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.950026 4930 manager.go:324] Recovery completed Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.961860 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.963970 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.964097 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.964226 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.965010 4930 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.965101 4930 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.965195 4930 state_mem.go:36] "Initialized new in-memory state store" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.967226 4930 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.969484 4930 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.969560 4930 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.969599 4930 kubelet.go:2335] "Starting kubelet main sync loop" Mar 13 09:12:39 crc kubenswrapper[4930]: E0313 09:12:39.969728 4930 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 13 09:12:39 crc kubenswrapper[4930]: W0313 09:12:39.970352 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:39 crc kubenswrapper[4930]: E0313 09:12:39.970519 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.997943 4930 policy_none.go:49] "None policy: Start" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.998982 4930 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 13 09:12:39 crc kubenswrapper[4930]: I0313 09:12:39.999018 4930 state_mem.go:35] "Initializing new in-memory state store" Mar 13 09:12:40 crc kubenswrapper[4930]: E0313 09:12:40.013253 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.052458 4930 manager.go:334] "Starting Device Plugin manager" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.053350 4930 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.053388 4930 server.go:79] "Starting device plugin registration server" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.053884 4930 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.053911 4930 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.054146 4930 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.054291 4930 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.054305 4930 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 13 09:12:40 crc kubenswrapper[4930]: E0313 09:12:40.066561 4930 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.071044 4930 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.071141 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.072090 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.072118 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.072128 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.072285 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.072794 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.072988 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.073272 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.073318 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.073332 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.073559 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.073680 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.073722 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.074646 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.074675 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.074687 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.074811 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075012 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075123 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075217 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075303 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075329 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075339 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075235 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075604 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075625 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075635 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075657 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075754 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.075997 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.076090 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.076625 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.076659 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.076670 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.076762 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.076784 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.076795 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.076829 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.076859 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.077511 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.077531 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.077540 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.077645 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.077664 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.077675 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: E0313 09:12:40.116212 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="400ms" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.137831 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.137893 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.137916 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.137957 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.137979 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.137998 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.138131 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.138154 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.138200 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.138242 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.138268 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.138299 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.138318 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.138336 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.138373 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.154402 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.155333 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.155389 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.155401 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.155458 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:12:40 crc kubenswrapper[4930]: E0313 09:12:40.156173 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.238:6443: connect: connection refused" node="crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.239892 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.239949 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.239978 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240003 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240024 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240043 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240065 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240086 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240107 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240110 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240127 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240152 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240159 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240202 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240243 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240246 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240265 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240276 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240271 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240283 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240298 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240307 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240318 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240354 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240400 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240394 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240422 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240491 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240515 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.240606 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.356480 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.357805 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.357889 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.357908 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.357943 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:12:40 crc kubenswrapper[4930]: E0313 09:12:40.358609 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.238:6443: connect: connection refused" node="crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.408215 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.417295 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.435325 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.451421 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.457315 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:40 crc kubenswrapper[4930]: W0313 09:12:40.465474 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-308f0c8b17fb13f675355fb28104d785a9acd33bbecd8cc7f2bd52346a0ea971 WatchSource:0}: Error finding container 308f0c8b17fb13f675355fb28104d785a9acd33bbecd8cc7f2bd52346a0ea971: Status 404 returned error can't find the container with id 308f0c8b17fb13f675355fb28104d785a9acd33bbecd8cc7f2bd52346a0ea971 Mar 13 09:12:40 crc kubenswrapper[4930]: W0313 09:12:40.470168 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-6d898bb18bf90cd8b34495bb43b32173cdc68965396eaa367482ca04cdbb8364 WatchSource:0}: Error finding container 6d898bb18bf90cd8b34495bb43b32173cdc68965396eaa367482ca04cdbb8364: Status 404 returned error can't find the container with id 6d898bb18bf90cd8b34495bb43b32173cdc68965396eaa367482ca04cdbb8364 Mar 13 09:12:40 crc kubenswrapper[4930]: W0313 09:12:40.474848 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-affc329c2dca384b36e141d2828075a86024bac503aa04608b8df51f065ef385 WatchSource:0}: Error finding container affc329c2dca384b36e141d2828075a86024bac503aa04608b8df51f065ef385: Status 404 returned error can't find the container with id affc329c2dca384b36e141d2828075a86024bac503aa04608b8df51f065ef385 Mar 13 09:12:40 crc kubenswrapper[4930]: W0313 09:12:40.480860 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-f24dc93c0cbcad499ff5edd6ce607f194334fb02499217b4424ef11f608e3f73 WatchSource:0}: Error finding container f24dc93c0cbcad499ff5edd6ce607f194334fb02499217b4424ef11f608e3f73: Status 404 returned error can't find the container with id f24dc93c0cbcad499ff5edd6ce607f194334fb02499217b4424ef11f608e3f73 Mar 13 09:12:40 crc kubenswrapper[4930]: W0313 09:12:40.482381 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-5f39dfe28483764496c35e8de33f9a9087067caf32d84680b91b84e25547d60e WatchSource:0}: Error finding container 5f39dfe28483764496c35e8de33f9a9087067caf32d84680b91b84e25547d60e: Status 404 returned error can't find the container with id 5f39dfe28483764496c35e8de33f9a9087067caf32d84680b91b84e25547d60e Mar 13 09:12:40 crc kubenswrapper[4930]: E0313 09:12:40.518107 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="800ms" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.758885 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.760181 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.760256 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.760274 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.760304 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:12:40 crc kubenswrapper[4930]: E0313 09:12:40.760796 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.238:6443: connect: connection refused" node="crc" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.909365 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:40 crc kubenswrapper[4930]: W0313 09:12:40.920285 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:40 crc kubenswrapper[4930]: E0313 09:12:40.920402 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Mar 13 09:12:40 crc kubenswrapper[4930]: W0313 09:12:40.964591 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:40 crc kubenswrapper[4930]: E0313 09:12:40.964681 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.976776 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"308f0c8b17fb13f675355fb28104d785a9acd33bbecd8cc7f2bd52346a0ea971"} Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.977485 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5f39dfe28483764496c35e8de33f9a9087067caf32d84680b91b84e25547d60e"} Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.978414 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f24dc93c0cbcad499ff5edd6ce607f194334fb02499217b4424ef11f608e3f73"} Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.979264 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"affc329c2dca384b36e141d2828075a86024bac503aa04608b8df51f065ef385"} Mar 13 09:12:40 crc kubenswrapper[4930]: I0313 09:12:40.979991 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"6d898bb18bf90cd8b34495bb43b32173cdc68965396eaa367482ca04cdbb8364"} Mar 13 09:12:41 crc kubenswrapper[4930]: W0313 09:12:41.023601 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:41 crc kubenswrapper[4930]: E0313 09:12:41.023678 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Mar 13 09:12:41 crc kubenswrapper[4930]: W0313 09:12:41.147138 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:41 crc kubenswrapper[4930]: E0313 09:12:41.147569 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Mar 13 09:12:41 crc kubenswrapper[4930]: E0313 09:12:41.319667 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="1.6s" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.561027 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.562560 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.562603 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.562620 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.562649 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:12:41 crc kubenswrapper[4930]: E0313 09:12:41.563199 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.238:6443: connect: connection refused" node="crc" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.909927 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.958509 4930 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 09:12:41 crc kubenswrapper[4930]: E0313 09:12:41.959751 4930 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.987639 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.988415 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"db0f40c73e898e6cabd0345ff22905e597ab8f27f13fce91b84916f6ac3169d7"} Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.988520 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"26e3c78121d34a9ec19c0081b3f32a8ea8851430cdd14a298eda588b224688b0"} Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.988552 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4d39002d4d8e00ebd6d5063c8502f97a674a06b5625e75e180605846cb9802dd"} Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.988580 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"e3dc535836d240a06323d71572218900609f93fcf435b117c948297f522800cc"} Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.989218 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.989285 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.989307 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.993185 4930 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="e48f4e8b192c1037cd477118f37919e9bc7126cf59b2a6095b0d684ce56c1f5f" exitCode=0 Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.993323 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.993338 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"e48f4e8b192c1037cd477118f37919e9bc7126cf59b2a6095b0d684ce56c1f5f"} Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.994835 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.994875 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.994892 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.997044 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d564c8034ceace0de6a72404a2167e81b62e93fba97cc5d8ff9b10cf95b16376" exitCode=0 Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.997159 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"d564c8034ceace0de6a72404a2167e81b62e93fba97cc5d8ff9b10cf95b16376"} Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.997264 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.998482 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.998516 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:41 crc kubenswrapper[4930]: I0313 09:12:41.998534 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.001392 4930 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="b41d6b9b5f204b817b33c875a43ee086935dbe7ee9a2186711c964e664dd3747" exitCode=0 Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.001780 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.001851 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"b41d6b9b5f204b817b33c875a43ee086935dbe7ee9a2186711c964e664dd3747"} Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.003513 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.003556 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.003560 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.003896 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.004282 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.004337 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.004355 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.006119 4930 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="9d5f68eeddfe8d4d1ab3e7b48c2ef39ef8508eff7db10db9c376bfd299748a10" exitCode=0 Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.006146 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"9d5f68eeddfe8d4d1ab3e7b48c2ef39ef8508eff7db10db9c376bfd299748a10"} Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.006215 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.008086 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.008116 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.008131 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:42 crc kubenswrapper[4930]: I0313 09:12:42.910057 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:42 crc kubenswrapper[4930]: E0313 09:12:42.924277 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="3.2s" Mar 13 09:12:42 crc kubenswrapper[4930]: W0313 09:12:42.943964 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:42 crc kubenswrapper[4930]: E0313 09:12:42.944073 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.008577 4930 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8803d0b3daaee963ed2e8c080ea50d8cf49fc82989a20f1a2d8fb7ca6ff58dc4" exitCode=0 Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.008636 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8803d0b3daaee963ed2e8c080ea50d8cf49fc82989a20f1a2d8fb7ca6ff58dc4"} Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.008669 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.009502 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.009528 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.009537 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.012558 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"219bb5acc41ca811e1ec6b0ac1189aa64207596c0cb72eda47d67e95545b364e"} Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.012577 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"08cbff57f3d4819c197f0d2d6ac06b8562a3f00891c3d9640de14101914de31e"} Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.012588 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"02a86f4198baffedc25ef8a9b11068c3e64ca1cddf968d18f4f2c26520c0db44"} Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.012641 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.013526 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.013542 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.013551 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.015509 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"41656d2695e180f85d0efe0f6d6aa8dbed7a3ffd3c3d9cc50c1bdc1f81b81dab"} Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.015528 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"eb903d30335ef135d24ba2cf2714eb31d42bab12a773363f702b6b416028de80"} Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.015537 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"89ff0f4d19a0b5edb6f7c03b5606947f56ff8850698cee5e92ddc35fae32b124"} Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.015545 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6762958ef4a5f1e05ea149905373f52a6282491fbf69009937361f5bb42b6fc4"} Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.016511 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.016840 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.017112 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"bb1e3de75ed03b7e44185eb7ab8df78138aeb01abc9c764309175f6df0433ade"} Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.019929 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.019946 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.019955 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.020369 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.020386 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.020393 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:43 crc kubenswrapper[4930]: W0313 09:12:43.143553 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.238:6443: connect: connection refused Mar 13 09:12:43 crc kubenswrapper[4930]: E0313 09:12:43.143616 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.238:6443: connect: connection refused" logger="UnhandledError" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.164042 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.165176 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.165210 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.165219 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:43 crc kubenswrapper[4930]: I0313 09:12:43.165243 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:12:43 crc kubenswrapper[4930]: E0313 09:12:43.165602 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.238:6443: connect: connection refused" node="crc" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.023425 4930 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ee28d9fdefa8719872401e6f13efa96dc437c066b69259cac6441ca789562d82" exitCode=0 Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.023507 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ee28d9fdefa8719872401e6f13efa96dc437c066b69259cac6441ca789562d82"} Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.023627 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.024937 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.025011 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.025031 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.030322 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f73e00650a255bc264c6a79db4f1dce25e349921d7617f18bfa341ee470ce29e"} Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.030394 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.030496 4930 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.030521 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.030549 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.036562 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.036608 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.036657 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.036684 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.036620 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.036718 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.036768 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.036779 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:44 crc kubenswrapper[4930]: I0313 09:12:44.036807 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.038879 4930 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.038925 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.038901 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"96ea430e466c4e320c6b22d50c9fe48e84718db3505be7a491385e44d57b9211"} Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.038990 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4cb8969c4cdc68eae6ac1012ad12af22227420e76f0e911ed6619a8f3f94e549"} Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.039021 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3eec3e243bf3e1a59f2646cb7a190d62db1c691d565840403efab57bd1b38938"} Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.039929 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.039965 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.039982 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.607972 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.608218 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.609808 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.609870 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:45 crc kubenswrapper[4930]: I0313 09:12:45.609891 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.014595 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.049089 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8709134701e8f8e00e390c2aebb8cd4e96331d4c6c4a6ebe8d87e55cb1c4c63f"} Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.049171 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bbfa5055bc9fdd59193e19cffceca72e5234571f70dd5da155c348358f130dba"} Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.049179 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.049123 4930 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.049363 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.050871 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.050912 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.050924 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.050974 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.050931 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.050998 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.157612 4930 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.365720 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.366746 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.366805 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.366828 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.366868 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.672167 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.672379 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.674224 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.674276 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.674301 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.683425 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:46 crc kubenswrapper[4930]: I0313 09:12:46.805156 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.051472 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.051539 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.051581 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.052955 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.053120 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.053021 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.053186 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.053198 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.053211 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.053218 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.053608 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.053745 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.126389 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 13 09:12:47 crc kubenswrapper[4930]: I0313 09:12:47.176893 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:48 crc kubenswrapper[4930]: I0313 09:12:48.054992 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:48 crc kubenswrapper[4930]: I0313 09:12:48.056308 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:48 crc kubenswrapper[4930]: I0313 09:12:48.057545 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:48 crc kubenswrapper[4930]: I0313 09:12:48.057590 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:48 crc kubenswrapper[4930]: I0313 09:12:48.057608 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:48 crc kubenswrapper[4930]: I0313 09:12:48.058648 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:48 crc kubenswrapper[4930]: I0313 09:12:48.058685 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:48 crc kubenswrapper[4930]: I0313 09:12:48.058702 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.149349 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.150590 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.152097 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.152167 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.152192 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.320184 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.320359 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.322409 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.322501 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.322525 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.683570 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.683805 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.685389 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.685478 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:49 crc kubenswrapper[4930]: I0313 09:12:49.685489 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:50 crc kubenswrapper[4930]: E0313 09:12:50.066757 4930 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 09:12:50 crc kubenswrapper[4930]: I0313 09:12:50.844128 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:50 crc kubenswrapper[4930]: I0313 09:12:50.844343 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:50 crc kubenswrapper[4930]: I0313 09:12:50.846160 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:50 crc kubenswrapper[4930]: I0313 09:12:50.846236 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:50 crc kubenswrapper[4930]: I0313 09:12:50.846275 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:53 crc kubenswrapper[4930]: W0313 09:12:53.645488 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 13 09:12:53 crc kubenswrapper[4930]: I0313 09:12:53.645620 4930 trace.go:236] Trace[1786700903]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (13-Mar-2026 09:12:43.644) (total time: 10001ms): Mar 13 09:12:53 crc kubenswrapper[4930]: Trace[1786700903]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:12:53.645) Mar 13 09:12:53 crc kubenswrapper[4930]: Trace[1786700903]: [10.001396572s] [10.001396572s] END Mar 13 09:12:53 crc kubenswrapper[4930]: E0313 09:12:53.645658 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 13 09:12:53 crc kubenswrapper[4930]: I0313 09:12:53.844140 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 09:12:53 crc kubenswrapper[4930]: I0313 09:12:53.844225 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 09:12:53 crc kubenswrapper[4930]: E0313 09:12:53.860613 4930 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:53Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c5baedf62bc90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.907359888 +0000 UTC m=+0.657274595,LastTimestamp:2026-03-13 09:12:39.907359888 +0000 UTC m=+0.657274595,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:12:53 crc kubenswrapper[4930]: I0313 09:12:53.863744 4930 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 09:12:53 crc kubenswrapper[4930]: I0313 09:12:53.863814 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 13 09:12:53 crc kubenswrapper[4930]: W0313 09:12:53.869367 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:53Z is after 2026-02-23T05:33:13Z Mar 13 09:12:53 crc kubenswrapper[4930]: E0313 09:12:53.869599 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:53Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 09:12:53 crc kubenswrapper[4930]: I0313 09:12:53.870375 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:53Z is after 2026-02-23T05:33:13Z Mar 13 09:12:53 crc kubenswrapper[4930]: W0313 09:12:53.871225 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:53Z is after 2026-02-23T05:33:13Z Mar 13 09:12:53 crc kubenswrapper[4930]: E0313 09:12:53.871353 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:53Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 09:12:53 crc kubenswrapper[4930]: E0313 09:12:53.872367 4930 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:53Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 09:12:53 crc kubenswrapper[4930]: E0313 09:12:53.878600 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:53Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 13 09:12:53 crc kubenswrapper[4930]: I0313 09:12:53.880544 4930 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 09:12:53 crc kubenswrapper[4930]: I0313 09:12:53.880720 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 13 09:12:53 crc kubenswrapper[4930]: E0313 09:12:53.880602 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:53Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 09:12:53 crc kubenswrapper[4930]: W0313 09:12:53.902132 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:53Z is after 2026-02-23T05:33:13Z Mar 13 09:12:53 crc kubenswrapper[4930]: E0313 09:12:53.902231 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:53Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 09:12:53 crc kubenswrapper[4930]: I0313 09:12:53.913333 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:53Z is after 2026-02-23T05:33:13Z Mar 13 09:12:54 crc kubenswrapper[4930]: I0313 09:12:54.072884 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 09:12:54 crc kubenswrapper[4930]: I0313 09:12:54.074658 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f73e00650a255bc264c6a79db4f1dce25e349921d7617f18bfa341ee470ce29e" exitCode=255 Mar 13 09:12:54 crc kubenswrapper[4930]: I0313 09:12:54.074711 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f73e00650a255bc264c6a79db4f1dce25e349921d7617f18bfa341ee470ce29e"} Mar 13 09:12:54 crc kubenswrapper[4930]: I0313 09:12:54.074819 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:54 crc kubenswrapper[4930]: I0313 09:12:54.075567 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:54 crc kubenswrapper[4930]: I0313 09:12:54.075595 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:54 crc kubenswrapper[4930]: I0313 09:12:54.075609 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:54 crc kubenswrapper[4930]: I0313 09:12:54.076103 4930 scope.go:117] "RemoveContainer" containerID="f73e00650a255bc264c6a79db4f1dce25e349921d7617f18bfa341ee470ce29e" Mar 13 09:12:54 crc kubenswrapper[4930]: I0313 09:12:54.914894 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:54Z is after 2026-02-23T05:33:13Z Mar 13 09:12:55 crc kubenswrapper[4930]: I0313 09:12:55.078219 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 09:12:55 crc kubenswrapper[4930]: I0313 09:12:55.080355 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0f557de8dca1e83b5456841eedf72617da6bbd7f501f895f16685263a0366cec"} Mar 13 09:12:55 crc kubenswrapper[4930]: I0313 09:12:55.080497 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:55 crc kubenswrapper[4930]: I0313 09:12:55.081329 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:55 crc kubenswrapper[4930]: I0313 09:12:55.081395 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:55 crc kubenswrapper[4930]: I0313 09:12:55.081414 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:55 crc kubenswrapper[4930]: I0313 09:12:55.912943 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:55Z is after 2026-02-23T05:33:13Z Mar 13 09:12:56 crc kubenswrapper[4930]: I0313 09:12:56.084940 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 09:12:56 crc kubenswrapper[4930]: I0313 09:12:56.085757 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 13 09:12:56 crc kubenswrapper[4930]: I0313 09:12:56.088069 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0f557de8dca1e83b5456841eedf72617da6bbd7f501f895f16685263a0366cec" exitCode=255 Mar 13 09:12:56 crc kubenswrapper[4930]: I0313 09:12:56.088133 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"0f557de8dca1e83b5456841eedf72617da6bbd7f501f895f16685263a0366cec"} Mar 13 09:12:56 crc kubenswrapper[4930]: I0313 09:12:56.088173 4930 scope.go:117] "RemoveContainer" containerID="f73e00650a255bc264c6a79db4f1dce25e349921d7617f18bfa341ee470ce29e" Mar 13 09:12:56 crc kubenswrapper[4930]: I0313 09:12:56.088374 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:56 crc kubenswrapper[4930]: I0313 09:12:56.089414 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:56 crc kubenswrapper[4930]: I0313 09:12:56.089671 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:56 crc kubenswrapper[4930]: I0313 09:12:56.090699 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:56 crc kubenswrapper[4930]: I0313 09:12:56.092371 4930 scope.go:117] "RemoveContainer" containerID="0f557de8dca1e83b5456841eedf72617da6bbd7f501f895f16685263a0366cec" Mar 13 09:12:56 crc kubenswrapper[4930]: E0313 09:12:56.092838 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:12:56 crc kubenswrapper[4930]: I0313 09:12:56.806104 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:56 crc kubenswrapper[4930]: I0313 09:12:56.913980 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:56Z is after 2026-02-23T05:33:13Z Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.092230 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.094550 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.095281 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.095321 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.095335 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.095980 4930 scope.go:117] "RemoveContainer" containerID="0f557de8dca1e83b5456841eedf72617da6bbd7f501f895f16685263a0366cec" Mar 13 09:12:57 crc kubenswrapper[4930]: E0313 09:12:57.096156 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.160643 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.160859 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.162286 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.162344 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.162363 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.175778 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.183985 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:57 crc kubenswrapper[4930]: W0313 09:12:57.665094 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:57Z is after 2026-02-23T05:33:13Z Mar 13 09:12:57 crc kubenswrapper[4930]: E0313 09:12:57.665514 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:57Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 09:12:57 crc kubenswrapper[4930]: W0313 09:12:57.804299 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:57Z is after 2026-02-23T05:33:13Z Mar 13 09:12:57 crc kubenswrapper[4930]: E0313 09:12:57.804362 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:57Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 09:12:57 crc kubenswrapper[4930]: I0313 09:12:57.913523 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:57Z is after 2026-02-23T05:33:13Z Mar 13 09:12:58 crc kubenswrapper[4930]: I0313 09:12:58.097809 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:58 crc kubenswrapper[4930]: I0313 09:12:58.097857 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:58 crc kubenswrapper[4930]: I0313 09:12:58.099559 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:58 crc kubenswrapper[4930]: I0313 09:12:58.099590 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:58 crc kubenswrapper[4930]: I0313 09:12:58.099601 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:58 crc kubenswrapper[4930]: I0313 09:12:58.101171 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:58 crc kubenswrapper[4930]: I0313 09:12:58.101400 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:58 crc kubenswrapper[4930]: I0313 09:12:58.101538 4930 scope.go:117] "RemoveContainer" containerID="0f557de8dca1e83b5456841eedf72617da6bbd7f501f895f16685263a0366cec" Mar 13 09:12:58 crc kubenswrapper[4930]: I0313 09:12:58.101840 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:58 crc kubenswrapper[4930]: E0313 09:12:58.103460 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:12:58 crc kubenswrapper[4930]: I0313 09:12:58.107816 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:58 crc kubenswrapper[4930]: I0313 09:12:58.914409 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:58Z is after 2026-02-23T05:33:13Z Mar 13 09:12:59 crc kubenswrapper[4930]: I0313 09:12:59.100358 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:59 crc kubenswrapper[4930]: I0313 09:12:59.101847 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:59 crc kubenswrapper[4930]: I0313 09:12:59.101906 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:59 crc kubenswrapper[4930]: I0313 09:12:59.101926 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:59 crc kubenswrapper[4930]: I0313 09:12:59.102802 4930 scope.go:117] "RemoveContainer" containerID="0f557de8dca1e83b5456841eedf72617da6bbd7f501f895f16685263a0366cec" Mar 13 09:12:59 crc kubenswrapper[4930]: E0313 09:12:59.103152 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:12:59 crc kubenswrapper[4930]: I0313 09:12:59.227046 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:12:59 crc kubenswrapper[4930]: I0313 09:12:59.328794 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:12:59 crc kubenswrapper[4930]: I0313 09:12:59.329012 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:12:59 crc kubenswrapper[4930]: I0313 09:12:59.330670 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:12:59 crc kubenswrapper[4930]: I0313 09:12:59.330726 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:12:59 crc kubenswrapper[4930]: I0313 09:12:59.330744 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:12:59 crc kubenswrapper[4930]: I0313 09:12:59.913822 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:12:59Z is after 2026-02-23T05:33:13Z Mar 13 09:13:00 crc kubenswrapper[4930]: E0313 09:13:00.066947 4930 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 09:13:00 crc kubenswrapper[4930]: I0313 09:13:00.102722 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:00 crc kubenswrapper[4930]: I0313 09:13:00.104841 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:00 crc kubenswrapper[4930]: I0313 09:13:00.104890 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:00 crc kubenswrapper[4930]: I0313 09:13:00.104907 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:00 crc kubenswrapper[4930]: I0313 09:13:00.105880 4930 scope.go:117] "RemoveContainer" containerID="0f557de8dca1e83b5456841eedf72617da6bbd7f501f895f16685263a0366cec" Mar 13 09:13:00 crc kubenswrapper[4930]: E0313 09:13:00.106178 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:13:00 crc kubenswrapper[4930]: I0313 09:13:00.281979 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:00 crc kubenswrapper[4930]: I0313 09:13:00.283406 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:00 crc kubenswrapper[4930]: I0313 09:13:00.283482 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:00 crc kubenswrapper[4930]: I0313 09:13:00.283504 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:00 crc kubenswrapper[4930]: I0313 09:13:00.283539 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:13:00 crc kubenswrapper[4930]: E0313 09:13:00.284163 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:00Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 09:13:00 crc kubenswrapper[4930]: E0313 09:13:00.288276 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:00Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 09:13:00 crc kubenswrapper[4930]: W0313 09:13:00.464033 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:00Z is after 2026-02-23T05:33:13Z Mar 13 09:13:00 crc kubenswrapper[4930]: E0313 09:13:00.464131 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:00Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 09:13:00 crc kubenswrapper[4930]: I0313 09:13:00.915019 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:00Z is after 2026-02-23T05:33:13Z Mar 13 09:13:01 crc kubenswrapper[4930]: W0313 09:13:01.872118 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:01Z is after 2026-02-23T05:33:13Z Mar 13 09:13:01 crc kubenswrapper[4930]: E0313 09:13:01.872220 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:01Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 09:13:01 crc kubenswrapper[4930]: I0313 09:13:01.914233 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:01Z is after 2026-02-23T05:33:13Z Mar 13 09:13:02 crc kubenswrapper[4930]: I0313 09:13:02.464825 4930 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 09:13:02 crc kubenswrapper[4930]: E0313 09:13:02.470704 4930 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:02Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 09:13:02 crc kubenswrapper[4930]: I0313 09:13:02.913947 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:02Z is after 2026-02-23T05:33:13Z Mar 13 09:13:03 crc kubenswrapper[4930]: I0313 09:13:03.845340 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 09:13:03 crc kubenswrapper[4930]: I0313 09:13:03.845481 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 09:13:03 crc kubenswrapper[4930]: E0313 09:13:03.867503 4930 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:03Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c5baedf62bc90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.907359888 +0000 UTC m=+0.657274595,LastTimestamp:2026-03-13 09:12:39.907359888 +0000 UTC m=+0.657274595,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:03 crc kubenswrapper[4930]: I0313 09:13:03.912578 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:03Z is after 2026-02-23T05:33:13Z Mar 13 09:13:04 crc kubenswrapper[4930]: I0313 09:13:04.914630 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:04Z is after 2026-02-23T05:33:13Z Mar 13 09:13:05 crc kubenswrapper[4930]: I0313 09:13:05.914222 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:05Z is after 2026-02-23T05:33:13Z Mar 13 09:13:06 crc kubenswrapper[4930]: I0313 09:13:06.913271 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:06Z is after 2026-02-23T05:33:13Z Mar 13 09:13:07 crc kubenswrapper[4930]: E0313 09:13:07.287174 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:07Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 09:13:07 crc kubenswrapper[4930]: I0313 09:13:07.289363 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:07 crc kubenswrapper[4930]: I0313 09:13:07.290588 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:07 crc kubenswrapper[4930]: I0313 09:13:07.290619 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:07 crc kubenswrapper[4930]: I0313 09:13:07.290628 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:07 crc kubenswrapper[4930]: I0313 09:13:07.290648 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:13:07 crc kubenswrapper[4930]: E0313 09:13:07.293062 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:07Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 09:13:07 crc kubenswrapper[4930]: I0313 09:13:07.913376 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:07Z is after 2026-02-23T05:33:13Z Mar 13 09:13:08 crc kubenswrapper[4930]: W0313 09:13:08.134288 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:08Z is after 2026-02-23T05:33:13Z Mar 13 09:13:08 crc kubenswrapper[4930]: E0313 09:13:08.134397 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:08Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 09:13:08 crc kubenswrapper[4930]: W0313 09:13:08.392531 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:08Z is after 2026-02-23T05:33:13Z Mar 13 09:13:08 crc kubenswrapper[4930]: E0313 09:13:08.392629 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:08Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 13 09:13:08 crc kubenswrapper[4930]: I0313 09:13:08.915059 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:08Z is after 2026-02-23T05:33:13Z Mar 13 09:13:09 crc kubenswrapper[4930]: I0313 09:13:09.911704 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:09Z is after 2026-02-23T05:33:13Z Mar 13 09:13:10 crc kubenswrapper[4930]: E0313 09:13:10.067479 4930 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 09:13:10 crc kubenswrapper[4930]: I0313 09:13:10.915468 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:10Z is after 2026-02-23T05:33:13Z Mar 13 09:13:11 crc kubenswrapper[4930]: I0313 09:13:11.914612 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:11Z is after 2026-02-23T05:33:13Z Mar 13 09:13:11 crc kubenswrapper[4930]: I0313 09:13:11.970268 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:11 crc kubenswrapper[4930]: I0313 09:13:11.971748 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:11 crc kubenswrapper[4930]: I0313 09:13:11.971801 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:11 crc kubenswrapper[4930]: I0313 09:13:11.971820 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:11 crc kubenswrapper[4930]: I0313 09:13:11.972680 4930 scope.go:117] "RemoveContainer" containerID="0f557de8dca1e83b5456841eedf72617da6bbd7f501f895f16685263a0366cec" Mar 13 09:13:12 crc kubenswrapper[4930]: I0313 09:13:12.450216 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:44566->192.168.126.11:10357: read: connection reset by peer" start-of-body= Mar 13 09:13:12 crc kubenswrapper[4930]: I0313 09:13:12.450279 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:44566->192.168.126.11:10357: read: connection reset by peer" Mar 13 09:13:12 crc kubenswrapper[4930]: I0313 09:13:12.450333 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:13:12 crc kubenswrapper[4930]: I0313 09:13:12.450495 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:12 crc kubenswrapper[4930]: I0313 09:13:12.451650 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:12 crc kubenswrapper[4930]: I0313 09:13:12.451687 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:12 crc kubenswrapper[4930]: I0313 09:13:12.451705 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:12 crc kubenswrapper[4930]: I0313 09:13:12.452324 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"4d39002d4d8e00ebd6d5063c8502f97a674a06b5625e75e180605846cb9802dd"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 13 09:13:12 crc kubenswrapper[4930]: I0313 09:13:12.452614 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://4d39002d4d8e00ebd6d5063c8502f97a674a06b5625e75e180605846cb9802dd" gracePeriod=30 Mar 13 09:13:12 crc kubenswrapper[4930]: I0313 09:13:12.915952 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:12Z is after 2026-02-23T05:33:13Z Mar 13 09:13:13 crc kubenswrapper[4930]: I0313 09:13:13.142871 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 09:13:13 crc kubenswrapper[4930]: I0313 09:13:13.144727 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"392290f4e1dc2e3c38d309d8a85af6ac299ad33c0bd9d9dc15b2614861ff5e5b"} Mar 13 09:13:13 crc kubenswrapper[4930]: I0313 09:13:13.145013 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:13 crc kubenswrapper[4930]: I0313 09:13:13.146384 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:13 crc kubenswrapper[4930]: I0313 09:13:13.146425 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:13 crc kubenswrapper[4930]: I0313 09:13:13.146465 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:13 crc kubenswrapper[4930]: I0313 09:13:13.148488 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 09:13:13 crc kubenswrapper[4930]: I0313 09:13:13.148892 4930 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="4d39002d4d8e00ebd6d5063c8502f97a674a06b5625e75e180605846cb9802dd" exitCode=255 Mar 13 09:13:13 crc kubenswrapper[4930]: I0313 09:13:13.148927 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"4d39002d4d8e00ebd6d5063c8502f97a674a06b5625e75e180605846cb9802dd"} Mar 13 09:13:13 crc kubenswrapper[4930]: E0313 09:13:13.876190 4930 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:13Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189c5baedf62bc90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.907359888 +0000 UTC m=+0.657274595,LastTimestamp:2026-03-13 09:12:39.907359888 +0000 UTC m=+0.657274595,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:13 crc kubenswrapper[4930]: I0313 09:13:13.913929 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:13Z is after 2026-02-23T05:33:13Z Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.155252 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.156679 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.160247 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="392290f4e1dc2e3c38d309d8a85af6ac299ad33c0bd9d9dc15b2614861ff5e5b" exitCode=255 Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.160366 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"392290f4e1dc2e3c38d309d8a85af6ac299ad33c0bd9d9dc15b2614861ff5e5b"} Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.160509 4930 scope.go:117] "RemoveContainer" containerID="0f557de8dca1e83b5456841eedf72617da6bbd7f501f895f16685263a0366cec" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.160657 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.162175 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.162234 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.162259 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.163349 4930 scope.go:117] "RemoveContainer" containerID="392290f4e1dc2e3c38d309d8a85af6ac299ad33c0bd9d9dc15b2614861ff5e5b" Mar 13 09:13:14 crc kubenswrapper[4930]: E0313 09:13:14.163760 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.167263 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.168161 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3c1ae64e52d3ab03f858c87b36072a1bfc27b68265840d9d0e660f2cbcebe324"} Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.168316 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.170003 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.170053 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.170073 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:14 crc kubenswrapper[4930]: E0313 09:13:14.292653 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:14Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.293698 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.295233 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.295290 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.295308 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.295346 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:13:14 crc kubenswrapper[4930]: E0313 09:13:14.300002 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:14Z is after 2026-02-23T05:33:13Z" node="crc" Mar 13 09:13:14 crc kubenswrapper[4930]: I0313 09:13:14.911527 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:14Z is after 2026-02-23T05:33:13Z Mar 13 09:13:15 crc kubenswrapper[4930]: I0313 09:13:15.173142 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 09:13:15 crc kubenswrapper[4930]: I0313 09:13:15.175950 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:15 crc kubenswrapper[4930]: I0313 09:13:15.176779 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:15 crc kubenswrapper[4930]: I0313 09:13:15.176838 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:15 crc kubenswrapper[4930]: I0313 09:13:15.176856 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:15 crc kubenswrapper[4930]: I0313 09:13:15.608149 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:13:15 crc kubenswrapper[4930]: I0313 09:13:15.914618 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:13:15Z is after 2026-02-23T05:33:13Z Mar 13 09:13:16 crc kubenswrapper[4930]: I0313 09:13:16.177339 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:16 crc kubenswrapper[4930]: I0313 09:13:16.178794 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:16 crc kubenswrapper[4930]: I0313 09:13:16.178828 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:16 crc kubenswrapper[4930]: I0313 09:13:16.178841 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:16 crc kubenswrapper[4930]: I0313 09:13:16.805952 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:13:16 crc kubenswrapper[4930]: I0313 09:13:16.806199 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:16 crc kubenswrapper[4930]: I0313 09:13:16.807761 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:16 crc kubenswrapper[4930]: I0313 09:13:16.807888 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:16 crc kubenswrapper[4930]: I0313 09:13:16.807909 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:16 crc kubenswrapper[4930]: I0313 09:13:16.808846 4930 scope.go:117] "RemoveContainer" containerID="392290f4e1dc2e3c38d309d8a85af6ac299ad33c0bd9d9dc15b2614861ff5e5b" Mar 13 09:13:16 crc kubenswrapper[4930]: E0313 09:13:16.809187 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:13:16 crc kubenswrapper[4930]: I0313 09:13:16.916848 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:17 crc kubenswrapper[4930]: I0313 09:13:17.916937 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:18 crc kubenswrapper[4930]: I0313 09:13:18.591831 4930 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 13 09:13:18 crc kubenswrapper[4930]: I0313 09:13:18.611007 4930 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 13 09:13:18 crc kubenswrapper[4930]: I0313 09:13:18.913035 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:19 crc kubenswrapper[4930]: I0313 09:13:19.227234 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:13:19 crc kubenswrapper[4930]: I0313 09:13:19.227417 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:19 crc kubenswrapper[4930]: I0313 09:13:19.228725 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:19 crc kubenswrapper[4930]: I0313 09:13:19.228775 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:19 crc kubenswrapper[4930]: I0313 09:13:19.228791 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:19 crc kubenswrapper[4930]: I0313 09:13:19.229464 4930 scope.go:117] "RemoveContainer" containerID="392290f4e1dc2e3c38d309d8a85af6ac299ad33c0bd9d9dc15b2614861ff5e5b" Mar 13 09:13:19 crc kubenswrapper[4930]: E0313 09:13:19.229678 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:13:19 crc kubenswrapper[4930]: W0313 09:13:19.435827 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 13 09:13:19 crc kubenswrapper[4930]: E0313 09:13:19.435887 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 13 09:13:19 crc kubenswrapper[4930]: I0313 09:13:19.916560 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:20 crc kubenswrapper[4930]: E0313 09:13:20.067591 4930 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 09:13:20 crc kubenswrapper[4930]: I0313 09:13:20.844741 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:13:20 crc kubenswrapper[4930]: I0313 09:13:20.845022 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:20 crc kubenswrapper[4930]: I0313 09:13:20.846969 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:20 crc kubenswrapper[4930]: I0313 09:13:20.847030 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:20 crc kubenswrapper[4930]: I0313 09:13:20.847048 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:20 crc kubenswrapper[4930]: I0313 09:13:20.914975 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:21 crc kubenswrapper[4930]: E0313 09:13:21.298873 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 09:13:21 crc kubenswrapper[4930]: I0313 09:13:21.301129 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:21 crc kubenswrapper[4930]: I0313 09:13:21.302294 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:21 crc kubenswrapper[4930]: I0313 09:13:21.302330 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:21 crc kubenswrapper[4930]: I0313 09:13:21.302342 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:21 crc kubenswrapper[4930]: I0313 09:13:21.302367 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:13:21 crc kubenswrapper[4930]: E0313 09:13:21.308396 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 09:13:21 crc kubenswrapper[4930]: I0313 09:13:21.914502 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:22 crc kubenswrapper[4930]: I0313 09:13:22.916182 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:23 crc kubenswrapper[4930]: I0313 09:13:23.845732 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 09:13:23 crc kubenswrapper[4930]: I0313 09:13:23.845825 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.884514 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baedf62bc90 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.907359888 +0000 UTC m=+0.657274595,LastTimestamp:2026-03-13 09:12:39.907359888 +0000 UTC m=+0.657274595,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.891359 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c4453c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964083516 +0000 UTC m=+0.713998203,LastTimestamp:2026-03-13 09:12:39.964083516 +0000 UTC m=+0.713998203,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.898058 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c6499c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964215708 +0000 UTC m=+0.714130395,LastTimestamp:2026-03-13 09:12:39.964215708 +0000 UTC m=+0.714130395,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.904950 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c80482 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.96432909 +0000 UTC m=+0.714243777,LastTimestamp:2026-03-13 09:12:39.96432909 +0000 UTC m=+0.714243777,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: I0313 09:13:23.911357 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.911480 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee8366dc6 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:40.055451078 +0000 UTC m=+0.805365765,LastTimestamp:2026-03-13 09:12:40.055451078 +0000 UTC m=+0.805365765,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.924831 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c4453c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c4453c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964083516 +0000 UTC m=+0.713998203,LastTimestamp:2026-03-13 09:12:40.072111165 +0000 UTC m=+0.822025842,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.933483 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c6499c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c6499c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964215708 +0000 UTC m=+0.714130395,LastTimestamp:2026-03-13 09:12:40.072124526 +0000 UTC m=+0.822039203,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.940319 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c80482\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c80482 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.96432909 +0000 UTC m=+0.714243777,LastTimestamp:2026-03-13 09:12:40.072133416 +0000 UTC m=+0.822048093,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.947012 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c4453c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c4453c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964083516 +0000 UTC m=+0.713998203,LastTimestamp:2026-03-13 09:12:40.073301612 +0000 UTC m=+0.823216299,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.953529 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c6499c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c6499c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964215708 +0000 UTC m=+0.714130395,LastTimestamp:2026-03-13 09:12:40.073326823 +0000 UTC m=+0.823241510,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.959416 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c80482\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c80482 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.96432909 +0000 UTC m=+0.714243777,LastTimestamp:2026-03-13 09:12:40.073339353 +0000 UTC m=+0.823254040,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.965613 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c4453c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c4453c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964083516 +0000 UTC m=+0.713998203,LastTimestamp:2026-03-13 09:12:40.074667412 +0000 UTC m=+0.824582109,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.971734 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c6499c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c6499c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964215708 +0000 UTC m=+0.714130395,LastTimestamp:2026-03-13 09:12:40.074682432 +0000 UTC m=+0.824597119,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.978199 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c80482\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c80482 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.96432909 +0000 UTC m=+0.714243777,LastTimestamp:2026-03-13 09:12:40.074693402 +0000 UTC m=+0.824608089,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.984308 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c4453c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c4453c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964083516 +0000 UTC m=+0.713998203,LastTimestamp:2026-03-13 09:12:40.075115478 +0000 UTC m=+0.825030165,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.990183 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c6499c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c6499c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964215708 +0000 UTC m=+0.714130395,LastTimestamp:2026-03-13 09:12:40.0752107 +0000 UTC m=+0.825125387,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:23 crc kubenswrapper[4930]: E0313 09:13:23.996371 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c80482\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c80482 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.96432909 +0000 UTC m=+0.714243777,LastTimestamp:2026-03-13 09:12:40.075303231 +0000 UTC m=+0.825217918,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.000909 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c4453c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c4453c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964083516 +0000 UTC m=+0.713998203,LastTimestamp:2026-03-13 09:12:40.075318141 +0000 UTC m=+0.825232828,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.007047 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c6499c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c6499c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964215708 +0000 UTC m=+0.714130395,LastTimestamp:2026-03-13 09:12:40.075336101 +0000 UTC m=+0.825250788,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.011195 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c80482\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c80482 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.96432909 +0000 UTC m=+0.714243777,LastTimestamp:2026-03-13 09:12:40.075345722 +0000 UTC m=+0.825260409,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.015160 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c4453c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c4453c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964083516 +0000 UTC m=+0.713998203,LastTimestamp:2026-03-13 09:12:40.075620266 +0000 UTC m=+0.825534953,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.020178 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c6499c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c6499c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964215708 +0000 UTC m=+0.714130395,LastTimestamp:2026-03-13 09:12:40.075631696 +0000 UTC m=+0.825546383,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.024709 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c80482\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c80482 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.96432909 +0000 UTC m=+0.714243777,LastTimestamp:2026-03-13 09:12:40.075640896 +0000 UTC m=+0.825555583,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.027995 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c4453c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c4453c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964083516 +0000 UTC m=+0.713998203,LastTimestamp:2026-03-13 09:12:40.07664573 +0000 UTC m=+0.826560407,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.033369 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189c5baee2c6499c\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189c5baee2c6499c default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:39.964215708 +0000 UTC m=+0.714130395,LastTimestamp:2026-03-13 09:12:40.07666721 +0000 UTC m=+0.826581887,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.039826 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5baf011a69fa openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:40.473045498 +0000 UTC m=+1.222960195,LastTimestamp:2026-03-13 09:12:40.473045498 +0000 UTC m=+1.222960195,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.043565 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5baf011c0a96 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:40.47315215 +0000 UTC m=+1.223066857,LastTimestamp:2026-03-13 09:12:40.47315215 +0000 UTC m=+1.223066857,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.047286 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf016c0112 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:40.478392594 +0000 UTC m=+1.228307271,LastTimestamp:2026-03-13 09:12:40.478392594 +0000 UTC m=+1.228307271,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.050849 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5baf02188b00 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:40.489700096 +0000 UTC m=+1.239614803,LastTimestamp:2026-03-13 09:12:40.489700096 +0000 UTC m=+1.239614803,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.055204 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf02192574 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:40.489739636 +0000 UTC m=+1.239654353,LastTimestamp:2026-03-13 09:12:40.489739636 +0000 UTC m=+1.239654353,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.059622 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf23f66c5b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.057889371 +0000 UTC m=+1.807804048,LastTimestamp:2026-03-13 09:12:41.057889371 +0000 UTC m=+1.807804048,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.062737 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5baf242a1884 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.06127578 +0000 UTC m=+1.811190457,LastTimestamp:2026-03-13 09:12:41.06127578 +0000 UTC m=+1.811190457,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.065949 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5baf243dff78 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.062580088 +0000 UTC m=+1.812494755,LastTimestamp:2026-03-13 09:12:41.062580088 +0000 UTC m=+1.812494755,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.069081 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5baf246292b0 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.064977072 +0000 UTC m=+1.814891749,LastTimestamp:2026-03-13 09:12:41.064977072 +0000 UTC m=+1.814891749,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.072150 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf246a1d39 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.065471289 +0000 UTC m=+1.815385966,LastTimestamp:2026-03-13 09:12:41.065471289 +0000 UTC m=+1.815385966,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.075379 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf24929ccb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.068125387 +0000 UTC m=+1.818040064,LastTimestamp:2026-03-13 09:12:41.068125387 +0000 UTC m=+1.818040064,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.079127 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf24b9661e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.070667294 +0000 UTC m=+1.820581971,LastTimestamp:2026-03-13 09:12:41.070667294 +0000 UTC m=+1.820581971,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.082167 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5baf24edf685 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.074112133 +0000 UTC m=+1.824026810,LastTimestamp:2026-03-13 09:12:41.074112133 +0000 UTC m=+1.824026810,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.085273 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5baf25644a47 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.081866823 +0000 UTC m=+1.831781500,LastTimestamp:2026-03-13 09:12:41.081866823 +0000 UTC m=+1.831781500,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.088205 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5baf2569b4a0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.082221728 +0000 UTC m=+1.832136395,LastTimestamp:2026-03-13 09:12:41.082221728 +0000 UTC m=+1.832136395,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.091882 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf262ec0be openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.095135422 +0000 UTC m=+1.845050099,LastTimestamp:2026-03-13 09:12:41.095135422 +0000 UTC m=+1.845050099,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.095589 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf38c92e73 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.407245939 +0000 UTC m=+2.157160656,LastTimestamp:2026-03-13 09:12:41.407245939 +0000 UTC m=+2.157160656,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.098645 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf39a6030d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.421718285 +0000 UTC m=+2.171633002,LastTimestamp:2026-03-13 09:12:41.421718285 +0000 UTC m=+2.171633002,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.101680 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf39bb78c3 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.423124675 +0000 UTC m=+2.173039392,LastTimestamp:2026-03-13 09:12:41.423124675 +0000 UTC m=+2.173039392,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.104976 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf485173be openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.667834814 +0000 UTC m=+2.417749531,LastTimestamp:2026-03-13 09:12:41.667834814 +0000 UTC m=+2.417749531,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.108236 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf493f6298 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.683427992 +0000 UTC m=+2.433342699,LastTimestamp:2026-03-13 09:12:41.683427992 +0000 UTC m=+2.433342699,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.111606 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf4954cb84 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.684831108 +0000 UTC m=+2.434745825,LastTimestamp:2026-03-13 09:12:41.684831108 +0000 UTC m=+2.434745825,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.114920 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf581c2b4c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.932778316 +0000 UTC m=+2.682693023,LastTimestamp:2026-03-13 09:12:41.932778316 +0000 UTC m=+2.682693023,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.118097 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf58e468fe openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.94590131 +0000 UTC m=+2.695816017,LastTimestamp:2026-03-13 09:12:41.94590131 +0000 UTC m=+2.695816017,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.119644 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5baf5c07f821 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.998563361 +0000 UTC m=+2.748478068,LastTimestamp:2026-03-13 09:12:41.998563361 +0000 UTC m=+2.748478068,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.122301 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf5c51741b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.003379227 +0000 UTC m=+2.753293914,LastTimestamp:2026-03-13 09:12:42.003379227 +0000 UTC m=+2.753293914,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.125673 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5baf5c6e1448 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.00525524 +0000 UTC m=+2.755169947,LastTimestamp:2026-03-13 09:12:42.00525524 +0000 UTC m=+2.755169947,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.129020 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5baf5cee497f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.013657471 +0000 UTC m=+2.763572168,LastTimestamp:2026-03-13 09:12:42.013657471 +0000 UTC m=+2.763572168,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.132113 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5baf69f4d43e openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.232190014 +0000 UTC m=+2.982104691,LastTimestamp:2026-03-13 09:12:42.232190014 +0000 UTC m=+2.982104691,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.135982 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5baf6a5f048e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.239149198 +0000 UTC m=+2.989063875,LastTimestamp:2026-03-13 09:12:42.239149198 +0000 UTC m=+2.989063875,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.139012 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5baf6a9474cc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.24265134 +0000 UTC m=+2.992566007,LastTimestamp:2026-03-13 09:12:42.24265134 +0000 UTC m=+2.992566007,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.142330 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf6a9e56a6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.243298982 +0000 UTC m=+2.993213659,LastTimestamp:2026-03-13 09:12:42.243298982 +0000 UTC m=+2.993213659,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.145606 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189c5baf6ac6367d openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.245912189 +0000 UTC m=+2.995826866,LastTimestamp:2026-03-13 09:12:42.245912189 +0000 UTC m=+2.995826866,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.148921 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5baf6ae91701 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.248197889 +0000 UTC m=+2.998112566,LastTimestamp:2026-03-13 09:12:42.248197889 +0000 UTC m=+2.998112566,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.153177 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5baf6af46a83 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.248940163 +0000 UTC m=+2.998854840,LastTimestamp:2026-03-13 09:12:42.248940163 +0000 UTC m=+2.998854840,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.156390 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf6b6af2ec openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.256708332 +0000 UTC m=+3.006623009,LastTimestamp:2026-03-13 09:12:42.256708332 +0000 UTC m=+3.006623009,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.160935 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf6b82819d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.258252189 +0000 UTC m=+3.008166866,LastTimestamp:2026-03-13 09:12:42.258252189 +0000 UTC m=+3.008166866,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.165418 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5baf6c66940c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.273199116 +0000 UTC m=+3.023113793,LastTimestamp:2026-03-13 09:12:42.273199116 +0000 UTC m=+3.023113793,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.169727 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5baf77ab8ba3 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.462268323 +0000 UTC m=+3.212183000,LastTimestamp:2026-03-13 09:12:42.462268323 +0000 UTC m=+3.212183000,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.173135 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf77ac1267 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.462302823 +0000 UTC m=+3.212217510,LastTimestamp:2026-03-13 09:12:42.462302823 +0000 UTC m=+3.212217510,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.177773 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5baf784e5f09 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.472939273 +0000 UTC m=+3.222853950,LastTimestamp:2026-03-13 09:12:42.472939273 +0000 UTC m=+3.222853950,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.182350 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5baf78643073 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.474369139 +0000 UTC m=+3.224283816,LastTimestamp:2026-03-13 09:12:42.474369139 +0000 UTC m=+3.224283816,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.186565 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf786f0096 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.475077782 +0000 UTC m=+3.224992469,LastTimestamp:2026-03-13 09:12:42.475077782 +0000 UTC m=+3.224992469,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.190328 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf78d37c0d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.481662989 +0000 UTC m=+3.231577676,LastTimestamp:2026-03-13 09:12:42.481662989 +0000 UTC m=+3.231577676,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.193906 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf8258fecc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.641407692 +0000 UTC m=+3.391322369,LastTimestamp:2026-03-13 09:12:42.641407692 +0000 UTC m=+3.391322369,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.197402 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5baf82e4a0a2 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.650558626 +0000 UTC m=+3.400473303,LastTimestamp:2026-03-13 09:12:42.650558626 +0000 UTC m=+3.400473303,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.201883 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf838176e1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.660837089 +0000 UTC m=+3.410751776,LastTimestamp:2026-03-13 09:12:42.660837089 +0000 UTC m=+3.410751776,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.205320 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf8394651a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.662077722 +0000 UTC m=+3.411992399,LastTimestamp:2026-03-13 09:12:42.662077722 +0000 UTC m=+3.411992399,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.208930 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189c5baf83a50c75 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.663169141 +0000 UTC m=+3.413083838,LastTimestamp:2026-03-13 09:12:42.663169141 +0000 UTC m=+3.413083838,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.213490 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf8d78ff3f openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.828054335 +0000 UTC m=+3.577969022,LastTimestamp:2026-03-13 09:12:42.828054335 +0000 UTC m=+3.577969022,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.217603 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf8e15fa05 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.838342149 +0000 UTC m=+3.588256846,LastTimestamp:2026-03-13 09:12:42.838342149 +0000 UTC m=+3.588256846,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.221084 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf8e246efc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.839289596 +0000 UTC m=+3.589204273,LastTimestamp:2026-03-13 09:12:42.839289596 +0000 UTC m=+3.589204273,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.222838 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5baf985e5e72 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:43.01085861 +0000 UTC m=+3.760773287,LastTimestamp:2026-03-13 09:12:43.01085861 +0000 UTC m=+3.760773287,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.225010 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf9a252846 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:43.040663622 +0000 UTC m=+3.790578309,LastTimestamp:2026-03-13 09:12:43.040663622 +0000 UTC m=+3.790578309,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.228331 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf9af45317 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:43.054240535 +0000 UTC m=+3.804155232,LastTimestamp:2026-03-13 09:12:43.054240535 +0000 UTC m=+3.804155232,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.233344 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bafa420cb9a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:43.208149914 +0000 UTC m=+3.958064581,LastTimestamp:2026-03-13 09:12:43.208149914 +0000 UTC m=+3.958064581,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.237684 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bafa4f54094 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:43.222073492 +0000 UTC m=+3.971988169,LastTimestamp:2026-03-13 09:12:43.222073492 +0000 UTC m=+3.971988169,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.241710 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bafd4ed5e20 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:44.026863136 +0000 UTC m=+4.776777843,LastTimestamp:2026-03-13 09:12:44.026863136 +0000 UTC m=+4.776777843,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.246287 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bafe35d87a6 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:44.269094822 +0000 UTC m=+5.019009549,LastTimestamp:2026-03-13 09:12:44.269094822 +0000 UTC m=+5.019009549,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.251249 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bafe4184ea5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:44.281335461 +0000 UTC m=+5.031250158,LastTimestamp:2026-03-13 09:12:44.281335461 +0000 UTC m=+5.031250158,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.255151 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bafe4337141 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:44.283113793 +0000 UTC m=+5.033028500,LastTimestamp:2026-03-13 09:12:44.283113793 +0000 UTC m=+5.033028500,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.259697 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5baff3dd533a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:44.545905466 +0000 UTC m=+5.295820153,LastTimestamp:2026-03-13 09:12:44.545905466 +0000 UTC m=+5.295820153,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.264352 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5baff4cf66c4 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:44.56177018 +0000 UTC m=+5.311684897,LastTimestamp:2026-03-13 09:12:44.56177018 +0000 UTC m=+5.311684897,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.268678 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5baff4df1cfc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:44.562799868 +0000 UTC m=+5.312714585,LastTimestamp:2026-03-13 09:12:44.562799868 +0000 UTC m=+5.312714585,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.272793 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bb003a26d2e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:44.810480942 +0000 UTC m=+5.560395619,LastTimestamp:2026-03-13 09:12:44.810480942 +0000 UTC m=+5.560395619,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.276537 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bb0046ee19b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:44.823880091 +0000 UTC m=+5.573794768,LastTimestamp:2026-03-13 09:12:44.823880091 +0000 UTC m=+5.573794768,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.279842 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bb0048554a9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:44.825351337 +0000 UTC m=+5.575266014,LastTimestamp:2026-03-13 09:12:44.825351337 +0000 UTC m=+5.575266014,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.283450 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bb013a53429 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:45.079098409 +0000 UTC m=+5.829013086,LastTimestamp:2026-03-13 09:12:45.079098409 +0000 UTC m=+5.829013086,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.287269 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bb0148024be openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:45.093446846 +0000 UTC m=+5.843361523,LastTimestamp:2026-03-13 09:12:45.093446846 +0000 UTC m=+5.843361523,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.290750 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bb01493d9e9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:45.094738409 +0000 UTC m=+5.844653116,LastTimestamp:2026-03-13 09:12:45.094738409 +0000 UTC m=+5.844653116,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.295733 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bb023e789bf openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:45.351881151 +0000 UTC m=+6.101795828,LastTimestamp:2026-03-13 09:12:45.351881151 +0000 UTC m=+6.101795828,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.300092 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189c5bb024d5fb77 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:45.367507831 +0000 UTC m=+6.117422548,LastTimestamp:2026-03-13 09:12:45.367507831 +0000 UTC m=+6.117422548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.306405 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 09:13:24 crc kubenswrapper[4930]: &Event{ObjectMeta:{kube-controller-manager-crc.189c5bb21e162243 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 09:13:24 crc kubenswrapper[4930]: body: Mar 13 09:13:24 crc kubenswrapper[4930]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:53.844206147 +0000 UTC m=+14.594120834,LastTimestamp:2026-03-13 09:12:53.844206147 +0000 UTC m=+14.594120834,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 09:13:24 crc kubenswrapper[4930]: > Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.310406 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5bb21e177ae1 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:53.844294369 +0000 UTC m=+14.594209066,LastTimestamp:2026-03-13 09:12:53.844294369 +0000 UTC m=+14.594209066,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.314058 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 13 09:13:24 crc kubenswrapper[4930]: &Event{ObjectMeta:{kube-apiserver-crc.189c5bb21f410530 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 13 09:13:24 crc kubenswrapper[4930]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 09:13:24 crc kubenswrapper[4930]: Mar 13 09:13:24 crc kubenswrapper[4930]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:53.863793968 +0000 UTC m=+14.613708655,LastTimestamp:2026-03-13 09:12:53.863793968 +0000 UTC m=+14.613708655,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 09:13:24 crc kubenswrapper[4930]: > Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.317530 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5bb21f41c13d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:53.863842109 +0000 UTC m=+14.613756796,LastTimestamp:2026-03-13 09:12:53.863842109 +0000 UTC m=+14.613756796,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.320904 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c5bb21f410530\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 13 09:13:24 crc kubenswrapper[4930]: &Event{ObjectMeta:{kube-apiserver-crc.189c5bb21f410530 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 13 09:13:24 crc kubenswrapper[4930]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 13 09:13:24 crc kubenswrapper[4930]: Mar 13 09:13:24 crc kubenswrapper[4930]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:53.863793968 +0000 UTC m=+14.613708655,LastTimestamp:2026-03-13 09:12:53.880687343 +0000 UTC m=+14.630602030,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 09:13:24 crc kubenswrapper[4930]: > Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.324262 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c5bb21f41c13d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5bb21f41c13d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:53.863842109 +0000 UTC m=+14.613756796,LastTimestamp:2026-03-13 09:12:53.880822317 +0000 UTC m=+14.630737004,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.328851 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c5baf8e246efc\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf8e246efc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:42.839289596 +0000 UTC m=+3.589204273,LastTimestamp:2026-03-13 09:12:54.077701387 +0000 UTC m=+14.827616064,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.332511 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c5baf9a252846\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf9a252846 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:43.040663622 +0000 UTC m=+3.790578309,LastTimestamp:2026-03-13 09:12:54.262789388 +0000 UTC m=+15.012704075,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.336088 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189c5baf9af45317\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189c5baf9af45317 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:43.054240535 +0000 UTC m=+3.804155232,LastTimestamp:2026-03-13 09:12:54.270937818 +0000 UTC m=+15.020852515,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.341672 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5bb21e162243\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 09:13:24 crc kubenswrapper[4930]: &Event{ObjectMeta:{kube-controller-manager-crc.189c5bb21e162243 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 09:13:24 crc kubenswrapper[4930]: body: Mar 13 09:13:24 crc kubenswrapper[4930]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:53.844206147 +0000 UTC m=+14.594120834,LastTimestamp:2026-03-13 09:13:03.845429618 +0000 UTC m=+24.595344335,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 09:13:24 crc kubenswrapper[4930]: > Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.346494 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5bb21e177ae1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5bb21e177ae1 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:53.844294369 +0000 UTC m=+14.594209066,LastTimestamp:2026-03-13 09:13:03.845555671 +0000 UTC m=+24.595470388,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.352458 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 09:13:24 crc kubenswrapper[4930]: &Event{ObjectMeta:{kube-controller-manager-crc.189c5bb67318057b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": read tcp 192.168.126.11:44566->192.168.126.11:10357: read: connection reset by peer Mar 13 09:13:24 crc kubenswrapper[4930]: body: Mar 13 09:13:24 crc kubenswrapper[4930]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:13:12.450262395 +0000 UTC m=+33.200177072,LastTimestamp:2026-03-13 09:13:12.450262395 +0000 UTC m=+33.200177072,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 09:13:24 crc kubenswrapper[4930]: > Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.356336 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5bb67318b378 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": read tcp 192.168.126.11:44566->192.168.126.11:10357: read: connection reset by peer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:13:12.450306936 +0000 UTC m=+33.200221613,LastTimestamp:2026-03-13 09:13:12.450306936 +0000 UTC m=+33.200221613,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.360685 4930 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5bb6733b7aa8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:13:12.452586152 +0000 UTC m=+33.202500869,LastTimestamp:2026-03-13 09:13:12.452586152 +0000 UTC m=+33.202500869,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.365466 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5baf24b9661e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf24b9661e openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.070667294 +0000 UTC m=+1.820581971,LastTimestamp:2026-03-13 09:13:12.974829892 +0000 UTC m=+33.724744609,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.370296 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5baf38c92e73\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf38c92e73 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.407245939 +0000 UTC m=+2.157160656,LastTimestamp:2026-03-13 09:13:13.204311574 +0000 UTC m=+33.954226251,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.375235 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5baf39a6030d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5baf39a6030d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:41.421718285 +0000 UTC m=+2.171633002,LastTimestamp:2026-03-13 09:13:13.214623578 +0000 UTC m=+33.964538255,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.382241 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5bb21e162243\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 13 09:13:24 crc kubenswrapper[4930]: &Event{ObjectMeta:{kube-controller-manager-crc.189c5bb21e162243 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 13 09:13:24 crc kubenswrapper[4930]: body: Mar 13 09:13:24 crc kubenswrapper[4930]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:53.844206147 +0000 UTC m=+14.594120834,LastTimestamp:2026-03-13 09:13:23.84579826 +0000 UTC m=+44.595712987,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 13 09:13:24 crc kubenswrapper[4930]: > Mar 13 09:13:24 crc kubenswrapper[4930]: E0313 09:13:24.387306 4930 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189c5bb21e177ae1\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189c5bb21e177ae1 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:12:53.844294369 +0000 UTC m=+14.594209066,LastTimestamp:2026-03-13 09:13:23.845873012 +0000 UTC m=+44.595787729,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:13:24 crc kubenswrapper[4930]: I0313 09:13:24.913972 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:25 crc kubenswrapper[4930]: W0313 09:13:25.176491 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:25 crc kubenswrapper[4930]: E0313 09:13:25.176570 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 13 09:13:25 crc kubenswrapper[4930]: W0313 09:13:25.441267 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 13 09:13:25 crc kubenswrapper[4930]: E0313 09:13:25.441331 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 13 09:13:25 crc kubenswrapper[4930]: I0313 09:13:25.916719 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:26 crc kubenswrapper[4930]: W0313 09:13:26.510801 4930 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 13 09:13:26 crc kubenswrapper[4930]: E0313 09:13:26.510854 4930 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 13 09:13:26 crc kubenswrapper[4930]: I0313 09:13:26.915468 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:27 crc kubenswrapper[4930]: I0313 09:13:27.913042 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:28 crc kubenswrapper[4930]: E0313 09:13:28.306703 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 09:13:28 crc kubenswrapper[4930]: I0313 09:13:28.308727 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:28 crc kubenswrapper[4930]: I0313 09:13:28.310098 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:28 crc kubenswrapper[4930]: I0313 09:13:28.310133 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:28 crc kubenswrapper[4930]: I0313 09:13:28.310145 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:28 crc kubenswrapper[4930]: I0313 09:13:28.310206 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:13:28 crc kubenswrapper[4930]: E0313 09:13:28.317499 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 09:13:28 crc kubenswrapper[4930]: I0313 09:13:28.913365 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:29 crc kubenswrapper[4930]: I0313 09:13:29.688974 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 09:13:29 crc kubenswrapper[4930]: I0313 09:13:29.689186 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:29 crc kubenswrapper[4930]: I0313 09:13:29.691008 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:29 crc kubenswrapper[4930]: I0313 09:13:29.691085 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:29 crc kubenswrapper[4930]: I0313 09:13:29.691099 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:29 crc kubenswrapper[4930]: I0313 09:13:29.914618 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:30 crc kubenswrapper[4930]: E0313 09:13:30.067903 4930 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 09:13:30 crc kubenswrapper[4930]: I0313 09:13:30.852191 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:13:30 crc kubenswrapper[4930]: I0313 09:13:30.852426 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:30 crc kubenswrapper[4930]: I0313 09:13:30.853855 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:30 crc kubenswrapper[4930]: I0313 09:13:30.853905 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:30 crc kubenswrapper[4930]: I0313 09:13:30.853925 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:30 crc kubenswrapper[4930]: I0313 09:13:30.859069 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:13:30 crc kubenswrapper[4930]: I0313 09:13:30.918541 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:31 crc kubenswrapper[4930]: I0313 09:13:31.215126 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:31 crc kubenswrapper[4930]: I0313 09:13:31.216683 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:31 crc kubenswrapper[4930]: I0313 09:13:31.216746 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:31 crc kubenswrapper[4930]: I0313 09:13:31.216768 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:31 crc kubenswrapper[4930]: I0313 09:13:31.916532 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:31 crc kubenswrapper[4930]: I0313 09:13:31.970713 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:31 crc kubenswrapper[4930]: I0313 09:13:31.972610 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:31 crc kubenswrapper[4930]: I0313 09:13:31.972694 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:31 crc kubenswrapper[4930]: I0313 09:13:31.972721 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:31 crc kubenswrapper[4930]: I0313 09:13:31.973702 4930 scope.go:117] "RemoveContainer" containerID="392290f4e1dc2e3c38d309d8a85af6ac299ad33c0bd9d9dc15b2614861ff5e5b" Mar 13 09:13:31 crc kubenswrapper[4930]: E0313 09:13:31.974099 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:13:32 crc kubenswrapper[4930]: I0313 09:13:32.915589 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:33 crc kubenswrapper[4930]: I0313 09:13:33.916406 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:34 crc kubenswrapper[4930]: I0313 09:13:34.914844 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:35 crc kubenswrapper[4930]: E0313 09:13:35.315248 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 09:13:35 crc kubenswrapper[4930]: I0313 09:13:35.318355 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:35 crc kubenswrapper[4930]: I0313 09:13:35.320133 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:35 crc kubenswrapper[4930]: I0313 09:13:35.320200 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:35 crc kubenswrapper[4930]: I0313 09:13:35.320225 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:35 crc kubenswrapper[4930]: I0313 09:13:35.320264 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:13:35 crc kubenswrapper[4930]: E0313 09:13:35.327178 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 09:13:35 crc kubenswrapper[4930]: I0313 09:13:35.914276 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:36 crc kubenswrapper[4930]: I0313 09:13:36.917110 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:37 crc kubenswrapper[4930]: I0313 09:13:37.915681 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:38 crc kubenswrapper[4930]: I0313 09:13:38.916346 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:39 crc kubenswrapper[4930]: I0313 09:13:39.913913 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:40 crc kubenswrapper[4930]: E0313 09:13:40.068794 4930 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 09:13:40 crc kubenswrapper[4930]: I0313 09:13:40.915541 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:41 crc kubenswrapper[4930]: I0313 09:13:41.915290 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:42 crc kubenswrapper[4930]: E0313 09:13:42.319741 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 13 09:13:42 crc kubenswrapper[4930]: I0313 09:13:42.327911 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:42 crc kubenswrapper[4930]: I0313 09:13:42.329475 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:42 crc kubenswrapper[4930]: I0313 09:13:42.329537 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:42 crc kubenswrapper[4930]: I0313 09:13:42.329553 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:42 crc kubenswrapper[4930]: I0313 09:13:42.329586 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:13:42 crc kubenswrapper[4930]: E0313 09:13:42.336924 4930 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 13 09:13:42 crc kubenswrapper[4930]: I0313 09:13:42.914554 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:43 crc kubenswrapper[4930]: I0313 09:13:43.915065 4930 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 13 09:13:44 crc kubenswrapper[4930]: I0313 09:13:44.467559 4930 csr.go:261] certificate signing request csr-g94g9 is approved, waiting to be issued Mar 13 09:13:44 crc kubenswrapper[4930]: I0313 09:13:44.476935 4930 csr.go:257] certificate signing request csr-g94g9 is issued Mar 13 09:13:44 crc kubenswrapper[4930]: I0313 09:13:44.558970 4930 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 13 09:13:44 crc kubenswrapper[4930]: I0313 09:13:44.746615 4930 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 13 09:13:45 crc kubenswrapper[4930]: I0313 09:13:45.478631 4930 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-19 07:36:45.717200829 +0000 UTC Mar 13 09:13:45 crc kubenswrapper[4930]: I0313 09:13:45.478678 4930 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6022h23m0.238525907s for next certificate rotation Mar 13 09:13:46 crc kubenswrapper[4930]: I0313 09:13:46.969884 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:46 crc kubenswrapper[4930]: I0313 09:13:46.971155 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:46 crc kubenswrapper[4930]: I0313 09:13:46.971213 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:46 crc kubenswrapper[4930]: I0313 09:13:46.971237 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:46 crc kubenswrapper[4930]: I0313 09:13:46.972316 4930 scope.go:117] "RemoveContainer" containerID="392290f4e1dc2e3c38d309d8a85af6ac299ad33c0bd9d9dc15b2614861ff5e5b" Mar 13 09:13:47 crc kubenswrapper[4930]: I0313 09:13:47.261052 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 09:13:47 crc kubenswrapper[4930]: I0313 09:13:47.262752 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1fc8d89dc87d04d45f03cdf66b778aceb7d3a920f5fd254edc0f7d5491c145bf"} Mar 13 09:13:47 crc kubenswrapper[4930]: I0313 09:13:47.262907 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:47 crc kubenswrapper[4930]: I0313 09:13:47.263828 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:47 crc kubenswrapper[4930]: I0313 09:13:47.263866 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:47 crc kubenswrapper[4930]: I0313 09:13:47.263875 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:48 crc kubenswrapper[4930]: I0313 09:13:48.266056 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 09:13:48 crc kubenswrapper[4930]: I0313 09:13:48.266497 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 13 09:13:48 crc kubenswrapper[4930]: I0313 09:13:48.267803 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1fc8d89dc87d04d45f03cdf66b778aceb7d3a920f5fd254edc0f7d5491c145bf" exitCode=255 Mar 13 09:13:48 crc kubenswrapper[4930]: I0313 09:13:48.267842 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1fc8d89dc87d04d45f03cdf66b778aceb7d3a920f5fd254edc0f7d5491c145bf"} Mar 13 09:13:48 crc kubenswrapper[4930]: I0313 09:13:48.267882 4930 scope.go:117] "RemoveContainer" containerID="392290f4e1dc2e3c38d309d8a85af6ac299ad33c0bd9d9dc15b2614861ff5e5b" Mar 13 09:13:48 crc kubenswrapper[4930]: I0313 09:13:48.268007 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:48 crc kubenswrapper[4930]: I0313 09:13:48.268824 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:48 crc kubenswrapper[4930]: I0313 09:13:48.268846 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:48 crc kubenswrapper[4930]: I0313 09:13:48.268854 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:48 crc kubenswrapper[4930]: I0313 09:13:48.269285 4930 scope.go:117] "RemoveContainer" containerID="1fc8d89dc87d04d45f03cdf66b778aceb7d3a920f5fd254edc0f7d5491c145bf" Mar 13 09:13:48 crc kubenswrapper[4930]: E0313 09:13:48.269415 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.227330 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.272578 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.276249 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.277134 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.277169 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.277180 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.277834 4930 scope.go:117] "RemoveContainer" containerID="1fc8d89dc87d04d45f03cdf66b778aceb7d3a920f5fd254edc0f7d5491c145bf" Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.278101 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.337829 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.339072 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.339105 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.339116 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.339210 4930 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.348132 4930 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.348403 4930 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.348426 4930 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.352148 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.352209 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.352223 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.352242 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.352256 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:13:49Z","lastTransitionTime":"2026-03-13T09:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.364504 4930 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"577d718c-b49d-4fb4-b481-cae7fc188388\\\",\\\"systemUUID\\\":\\\"f35d26c2-9d58-43d9-b3d1-f4e2250d2ace\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.373003 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.373076 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.373090 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.373129 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.373141 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:13:49Z","lastTransitionTime":"2026-03-13T09:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.386297 4930 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"577d718c-b49d-4fb4-b481-cae7fc188388\\\",\\\"systemUUID\\\":\\\"f35d26c2-9d58-43d9-b3d1-f4e2250d2ace\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.394670 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.394699 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.394710 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.394722 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.394731 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:13:49Z","lastTransitionTime":"2026-03-13T09:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.408445 4930 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"577d718c-b49d-4fb4-b481-cae7fc188388\\\",\\\"systemUUID\\\":\\\"f35d26c2-9d58-43d9-b3d1-f4e2250d2ace\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.417913 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.417959 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.417971 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.417988 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:13:49 crc kubenswrapper[4930]: I0313 09:13:49.418001 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:13:49Z","lastTransitionTime":"2026-03-13T09:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.428292 4930 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"577d718c-b49d-4fb4-b481-cae7fc188388\\\",\\\"systemUUID\\\":\\\"f35d26c2-9d58-43d9-b3d1-f4e2250d2ace\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.428464 4930 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.428492 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.529583 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.630537 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.730832 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.831522 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:49 crc kubenswrapper[4930]: E0313 09:13:49.931957 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:50 crc kubenswrapper[4930]: E0313 09:13:50.032241 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:50 crc kubenswrapper[4930]: E0313 09:13:50.068967 4930 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 09:13:50 crc kubenswrapper[4930]: E0313 09:13:50.133109 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:50 crc kubenswrapper[4930]: E0313 09:13:50.233237 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:50 crc kubenswrapper[4930]: E0313 09:13:50.334488 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:50 crc kubenswrapper[4930]: E0313 09:13:50.435047 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:50 crc kubenswrapper[4930]: E0313 09:13:50.535935 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:50 crc kubenswrapper[4930]: E0313 09:13:50.636980 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:50 crc kubenswrapper[4930]: E0313 09:13:50.737693 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:50 crc kubenswrapper[4930]: E0313 09:13:50.838287 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:50 crc kubenswrapper[4930]: E0313 09:13:50.938776 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:51 crc kubenswrapper[4930]: E0313 09:13:51.039509 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:51 crc kubenswrapper[4930]: E0313 09:13:51.140517 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:51 crc kubenswrapper[4930]: E0313 09:13:51.240876 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:51 crc kubenswrapper[4930]: E0313 09:13:51.340976 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:51 crc kubenswrapper[4930]: E0313 09:13:51.441938 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:51 crc kubenswrapper[4930]: E0313 09:13:51.542455 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:51 crc kubenswrapper[4930]: E0313 09:13:51.643508 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:51 crc kubenswrapper[4930]: E0313 09:13:51.744517 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:51 crc kubenswrapper[4930]: E0313 09:13:51.844854 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:51 crc kubenswrapper[4930]: E0313 09:13:51.945389 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:52 crc kubenswrapper[4930]: E0313 09:13:52.045798 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:52 crc kubenswrapper[4930]: E0313 09:13:52.146588 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:52 crc kubenswrapper[4930]: E0313 09:13:52.246967 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:52 crc kubenswrapper[4930]: E0313 09:13:52.348282 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:52 crc kubenswrapper[4930]: E0313 09:13:52.449390 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:52 crc kubenswrapper[4930]: E0313 09:13:52.550514 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:52 crc kubenswrapper[4930]: E0313 09:13:52.651411 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:52 crc kubenswrapper[4930]: E0313 09:13:52.752228 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:52 crc kubenswrapper[4930]: E0313 09:13:52.852936 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:52 crc kubenswrapper[4930]: E0313 09:13:52.953599 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:53 crc kubenswrapper[4930]: E0313 09:13:53.053977 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:53 crc kubenswrapper[4930]: E0313 09:13:53.155151 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:53 crc kubenswrapper[4930]: E0313 09:13:53.256111 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:53 crc kubenswrapper[4930]: E0313 09:13:53.356427 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:53 crc kubenswrapper[4930]: E0313 09:13:53.456753 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:53 crc kubenswrapper[4930]: E0313 09:13:53.557954 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:53 crc kubenswrapper[4930]: E0313 09:13:53.659061 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:53 crc kubenswrapper[4930]: E0313 09:13:53.760112 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:53 crc kubenswrapper[4930]: E0313 09:13:53.860652 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:53 crc kubenswrapper[4930]: E0313 09:13:53.961565 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:54 crc kubenswrapper[4930]: E0313 09:13:54.062400 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:54 crc kubenswrapper[4930]: E0313 09:13:54.162599 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:54 crc kubenswrapper[4930]: E0313 09:13:54.263671 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:54 crc kubenswrapper[4930]: E0313 09:13:54.364134 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:54 crc kubenswrapper[4930]: E0313 09:13:54.465293 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:54 crc kubenswrapper[4930]: E0313 09:13:54.566202 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:54 crc kubenswrapper[4930]: I0313 09:13:54.626081 4930 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 13 09:13:54 crc kubenswrapper[4930]: E0313 09:13:54.667165 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:54 crc kubenswrapper[4930]: E0313 09:13:54.767411 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:54 crc kubenswrapper[4930]: E0313 09:13:54.867875 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:54 crc kubenswrapper[4930]: E0313 09:13:54.969063 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:55 crc kubenswrapper[4930]: E0313 09:13:55.070076 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:55 crc kubenswrapper[4930]: E0313 09:13:55.171130 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:55 crc kubenswrapper[4930]: E0313 09:13:55.272120 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:55 crc kubenswrapper[4930]: E0313 09:13:55.372665 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:55 crc kubenswrapper[4930]: E0313 09:13:55.473801 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:55 crc kubenswrapper[4930]: E0313 09:13:55.574966 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:55 crc kubenswrapper[4930]: E0313 09:13:55.675149 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:55 crc kubenswrapper[4930]: E0313 09:13:55.776114 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:55 crc kubenswrapper[4930]: E0313 09:13:55.877055 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:55 crc kubenswrapper[4930]: E0313 09:13:55.977764 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:56 crc kubenswrapper[4930]: E0313 09:13:56.078787 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:56 crc kubenswrapper[4930]: E0313 09:13:56.179779 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:56 crc kubenswrapper[4930]: E0313 09:13:56.280209 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:56 crc kubenswrapper[4930]: E0313 09:13:56.381115 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:56 crc kubenswrapper[4930]: E0313 09:13:56.481393 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:56 crc kubenswrapper[4930]: E0313 09:13:56.582525 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:56 crc kubenswrapper[4930]: I0313 09:13:56.584687 4930 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 13 09:13:56 crc kubenswrapper[4930]: E0313 09:13:56.682740 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:56 crc kubenswrapper[4930]: E0313 09:13:56.783218 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:56 crc kubenswrapper[4930]: I0313 09:13:56.805410 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:13:56 crc kubenswrapper[4930]: I0313 09:13:56.805598 4930 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 13 09:13:56 crc kubenswrapper[4930]: I0313 09:13:56.806913 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:56 crc kubenswrapper[4930]: I0313 09:13:56.806964 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:56 crc kubenswrapper[4930]: I0313 09:13:56.806982 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:56 crc kubenswrapper[4930]: I0313 09:13:56.807819 4930 scope.go:117] "RemoveContainer" containerID="1fc8d89dc87d04d45f03cdf66b778aceb7d3a920f5fd254edc0f7d5491c145bf" Mar 13 09:13:56 crc kubenswrapper[4930]: E0313 09:13:56.808109 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:13:56 crc kubenswrapper[4930]: E0313 09:13:56.883674 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:56 crc kubenswrapper[4930]: E0313 09:13:56.984023 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:57 crc kubenswrapper[4930]: E0313 09:13:57.084820 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:57 crc kubenswrapper[4930]: E0313 09:13:57.185909 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:57 crc kubenswrapper[4930]: E0313 09:13:57.286594 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:57 crc kubenswrapper[4930]: E0313 09:13:57.387378 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:57 crc kubenswrapper[4930]: E0313 09:13:57.487987 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:57 crc kubenswrapper[4930]: E0313 09:13:57.588326 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:57 crc kubenswrapper[4930]: E0313 09:13:57.688464 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:57 crc kubenswrapper[4930]: E0313 09:13:57.789490 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:57 crc kubenswrapper[4930]: E0313 09:13:57.889933 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:57 crc kubenswrapper[4930]: E0313 09:13:57.990389 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:58 crc kubenswrapper[4930]: E0313 09:13:58.091372 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:58 crc kubenswrapper[4930]: E0313 09:13:58.191580 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:58 crc kubenswrapper[4930]: E0313 09:13:58.292219 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:58 crc kubenswrapper[4930]: E0313 09:13:58.393041 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:58 crc kubenswrapper[4930]: E0313 09:13:58.493358 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:58 crc kubenswrapper[4930]: E0313 09:13:58.594274 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:58 crc kubenswrapper[4930]: E0313 09:13:58.694576 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:58 crc kubenswrapper[4930]: I0313 09:13:58.783945 4930 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 13 09:13:58 crc kubenswrapper[4930]: E0313 09:13:58.795149 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:58 crc kubenswrapper[4930]: E0313 09:13:58.895601 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:58 crc kubenswrapper[4930]: E0313 09:13:58.996085 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.097219 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.198206 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.298777 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.399961 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.500159 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.601311 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.702193 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.796048 4930 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.800716 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.800856 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.800926 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.801002 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.801058 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:13:59Z","lastTransitionTime":"2026-03-13T09:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.814537 4930 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"577d718c-b49d-4fb4-b481-cae7fc188388\\\",\\\"systemUUID\\\":\\\"f35d26c2-9d58-43d9-b3d1-f4e2250d2ace\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.818355 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.818389 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.818398 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.818413 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.818421 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:13:59Z","lastTransitionTime":"2026-03-13T09:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.831656 4930 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"577d718c-b49d-4fb4-b481-cae7fc188388\\\",\\\"systemUUID\\\":\\\"f35d26c2-9d58-43d9-b3d1-f4e2250d2ace\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.835799 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.835928 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.835992 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.836062 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.836124 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:13:59Z","lastTransitionTime":"2026-03-13T09:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.849203 4930 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"577d718c-b49d-4fb4-b481-cae7fc188388\\\",\\\"systemUUID\\\":\\\"f35d26c2-9d58-43d9-b3d1-f4e2250d2ace\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.853895 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.853942 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.853956 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.853976 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:13:59 crc kubenswrapper[4930]: I0313 09:13:59.853988 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:13:59Z","lastTransitionTime":"2026-03-13T09:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.867141 4930 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-13T09:13:59Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"577d718c-b49d-4fb4-b481-cae7fc188388\\\",\\\"systemUUID\\\":\\\"f35d26c2-9d58-43d9-b3d1-f4e2250d2ace\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.867309 4930 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.867341 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:13:59 crc kubenswrapper[4930]: E0313 09:13:59.967514 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:00 crc kubenswrapper[4930]: E0313 09:14:00.067601 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:00 crc kubenswrapper[4930]: E0313 09:14:00.069864 4930 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 13 09:14:00 crc kubenswrapper[4930]: E0313 09:14:00.168149 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:00 crc kubenswrapper[4930]: E0313 09:14:00.268938 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:00 crc kubenswrapper[4930]: E0313 09:14:00.369721 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:00 crc kubenswrapper[4930]: E0313 09:14:00.470493 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:00 crc kubenswrapper[4930]: E0313 09:14:00.571174 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:00 crc kubenswrapper[4930]: E0313 09:14:00.672217 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:00 crc kubenswrapper[4930]: E0313 09:14:00.772744 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:00 crc kubenswrapper[4930]: E0313 09:14:00.873833 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:00 crc kubenswrapper[4930]: E0313 09:14:00.974987 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:01 crc kubenswrapper[4930]: E0313 09:14:01.075289 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:01 crc kubenswrapper[4930]: E0313 09:14:01.175844 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:01 crc kubenswrapper[4930]: E0313 09:14:01.276487 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:01 crc kubenswrapper[4930]: E0313 09:14:01.377610 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:01 crc kubenswrapper[4930]: E0313 09:14:01.477882 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:01 crc kubenswrapper[4930]: E0313 09:14:01.578709 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:01 crc kubenswrapper[4930]: E0313 09:14:01.678854 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:01 crc kubenswrapper[4930]: E0313 09:14:01.779270 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:01 crc kubenswrapper[4930]: E0313 09:14:01.879931 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:01 crc kubenswrapper[4930]: E0313 09:14:01.981034 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:02 crc kubenswrapper[4930]: E0313 09:14:02.081622 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:02 crc kubenswrapper[4930]: E0313 09:14:02.182704 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:02 crc kubenswrapper[4930]: E0313 09:14:02.283494 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:02 crc kubenswrapper[4930]: E0313 09:14:02.384789 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:02 crc kubenswrapper[4930]: E0313 09:14:02.485780 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:02 crc kubenswrapper[4930]: E0313 09:14:02.586498 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:02 crc kubenswrapper[4930]: E0313 09:14:02.687646 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:02 crc kubenswrapper[4930]: E0313 09:14:02.788646 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:02 crc kubenswrapper[4930]: E0313 09:14:02.889074 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:02 crc kubenswrapper[4930]: E0313 09:14:02.989693 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:03 crc kubenswrapper[4930]: E0313 09:14:03.090665 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:03 crc kubenswrapper[4930]: E0313 09:14:03.191447 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:03 crc kubenswrapper[4930]: E0313 09:14:03.292563 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:03 crc kubenswrapper[4930]: E0313 09:14:03.394241 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:03 crc kubenswrapper[4930]: E0313 09:14:03.495815 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:03 crc kubenswrapper[4930]: E0313 09:14:03.596767 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:03 crc kubenswrapper[4930]: E0313 09:14:03.697931 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:03 crc kubenswrapper[4930]: E0313 09:14:03.798363 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:03 crc kubenswrapper[4930]: E0313 09:14:03.898677 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:03 crc kubenswrapper[4930]: E0313 09:14:03.999041 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:04 crc kubenswrapper[4930]: E0313 09:14:04.099534 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:04 crc kubenswrapper[4930]: E0313 09:14:04.200272 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:04 crc kubenswrapper[4930]: E0313 09:14:04.301306 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:04 crc kubenswrapper[4930]: E0313 09:14:04.402233 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:04 crc kubenswrapper[4930]: E0313 09:14:04.503343 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:04 crc kubenswrapper[4930]: E0313 09:14:04.604380 4930 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.684470 4930 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.706236 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.706276 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.706291 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.706314 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.706363 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:04Z","lastTransitionTime":"2026-03-13T09:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.808508 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.808536 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.808552 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.808569 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.808579 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:04Z","lastTransitionTime":"2026-03-13T09:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.910637 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.910971 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.911096 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.911219 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:04 crc kubenswrapper[4930]: I0313 09:14:04.911343 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:04Z","lastTransitionTime":"2026-03-13T09:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.015414 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.015479 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.015494 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.015512 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.015527 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:05Z","lastTransitionTime":"2026-03-13T09:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.117969 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.118293 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.118465 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.118697 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.118843 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:05Z","lastTransitionTime":"2026-03-13T09:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.196745 4930 apiserver.go:52] "Watching apiserver" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.202798 4930 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.203274 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf"] Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.203820 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.203996 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.204095 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.204364 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.204505 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.204885 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.205799 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.205968 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.206317 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.206427 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.205915 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.206676 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.206869 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.208081 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.208160 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.209608 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.211152 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.212536 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.213855 4930 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.221097 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.221125 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.221133 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.221146 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.221177 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:05Z","lastTransitionTime":"2026-03-13T09:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.227671 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.248037 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.259415 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8dee6b1-215b-445a-9580-c3858a84e940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1e3de75ed03b7e44185eb7ab8df78138aeb01abc9c764309175f6df0433ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T09:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41d6b9b5f204b817b33c875a43ee086935dbe7ee9a2186711c964e664dd3747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b41d6b9b5f204b817b33c875a43ee086935dbe7ee9a2186711c964e664dd3747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T09:12:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T09:12:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T09:12:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.275816 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.293214 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295788 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295815 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295832 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295847 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295863 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295877 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295891 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295906 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295922 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295938 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295955 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295974 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.295992 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296007 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296021 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296038 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296055 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296073 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296088 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296106 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296122 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296137 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296151 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296166 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296182 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296196 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296210 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296223 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296238 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296253 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296268 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296285 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296302 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296317 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296334 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296349 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296363 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296379 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296394 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296442 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296463 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296477 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296511 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296527 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296542 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296558 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296612 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296636 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296676 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296693 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296707 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296722 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296757 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.296808 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297012 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297030 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297046 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297081 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297097 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297111 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297126 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297106 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297163 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297180 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297194 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297325 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297349 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297367 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297385 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297418 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297465 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297485 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297487 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297501 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297533 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297552 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297568 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297584 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297582 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297617 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297634 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297648 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297650 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297688 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297706 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297723 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297739 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297783 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297800 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297816 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297833 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297838 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297849 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297865 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297882 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297898 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297890 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297914 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297994 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297982 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298023 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298067 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298087 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298147 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298204 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298265 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298323 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298348 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298377 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298405 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298429 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298525 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298578 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298635 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298686 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298737 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298785 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298834 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298885 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298938 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298996 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299049 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299101 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299155 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299204 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299250 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299299 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299350 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299408 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299494 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299553 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299596 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299642 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299690 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299743 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299795 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299842 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299895 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299949 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300010 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300064 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300116 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300171 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300221 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300271 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300324 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300376 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300427 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300520 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300576 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300633 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300690 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300743 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300798 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300851 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300912 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300969 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301045 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301110 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301165 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301233 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301287 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301343 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301400 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301502 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301555 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301608 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301664 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301725 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301812 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301874 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301933 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301988 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.302046 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.308590 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298426 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298532 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298595 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298610 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298628 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298688 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298801 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298838 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298938 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298943 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.298986 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299069 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.297115 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299654 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299786 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299787 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.299851 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.300644 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301154 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301188 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301224 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301311 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301789 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301837 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.301919 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.302069 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.302103 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:05.802074383 +0000 UTC m=+86.551989100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309511 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309551 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309552 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309579 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309612 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309637 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309664 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309672 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309690 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309716 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309743 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309744 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309793 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309919 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309988 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310065 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310124 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310175 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310213 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310253 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310301 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310356 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310332 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310394 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310443 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310473 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310519 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310574 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310628 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310680 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310716 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310752 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310774 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310792 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310842 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310848 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310891 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310935 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310972 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311222 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311271 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311478 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311548 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311633 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311698 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311758 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311820 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311877 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311939 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311996 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.314216 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.314382 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.314546 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.314688 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.314814 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.315007 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.315142 4930 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.315521 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.316610 4930 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.318701 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.318770 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.318806 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.319072 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.319218 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.319268 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.319316 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.319629 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.319682 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.319713 4930 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.319759 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320033 4930 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320075 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320107 4930 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320149 4930 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320376 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320400 4930 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320422 4930 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320499 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320680 4930 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320712 4930 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320741 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320772 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321020 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321063 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321102 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321135 4930 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321331 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321356 4930 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321385 4930 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321409 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321635 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321663 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321693 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321908 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321944 4930 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.321990 4930 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.322026 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.322200 4930 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.322231 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.322275 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.322308 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.322490 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.322525 4930 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.334151 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.310897 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.336049 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.337059 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.337090 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.336215 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311122 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.302145 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.302227 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.302804 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.302903 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.303096 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.303044 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.303334 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.303389 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.303413 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.303512 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.303531 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.303487 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.304096 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.304187 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.304461 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.304480 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.304971 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.305310 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.305374 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.305554 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.305198 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.305855 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.305926 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.306084 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.306602 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.306607 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.306851 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.306911 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.308276 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.308574 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.308666 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.308784 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.308802 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.308820 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.308964 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.308989 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.309112 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311175 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311221 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311239 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311319 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311504 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311512 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311732 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.302130 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311865 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.311739 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.312007 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.312134 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.312258 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.312513 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.312925 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.312974 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.313094 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.313113 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.313141 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.313207 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.313382 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.313473 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.313844 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.314229 4930 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.314629 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.314653 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.314963 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.337928 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.314989 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.315756 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.315833 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.315875 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.316319 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.316441 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.316801 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.316859 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.316959 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.317138 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.317358 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.338211 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.318129 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.317865 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.318248 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.318658 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.318932 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.319047 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.319209 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.319466 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.319724 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.320739 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.323901 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.324298 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.326006 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.326134 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.326178 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.326340 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.335264 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.335552 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.333564 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.335710 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.336654 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.338002 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.338204 4930 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.338364 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.324638 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.338117 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.339034 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.341859 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.342005 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.342611 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.342640 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.343245 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.343304 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.343825 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.343300 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.344991 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.345198 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.345814 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.346064 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.346788 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.346973 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.347852 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.347952 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.348216 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:05.838412996 +0000 UTC m=+86.588327673 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.348286 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.348378 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:05.848346874 +0000 UTC m=+86.598261591 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.348854 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.349477 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.350154 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.352661 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.353362 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.353389 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.353420 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.353479 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:05Z","lastTransitionTime":"2026-03-13T09:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.348495 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.344198 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.349588 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.353883 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.353912 4930 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.349605 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.349906 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.350214 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.350312 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.350320 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.350730 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.350956 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.352959 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.354801 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.354828 4930 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.354944 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:05.854915025 +0000 UTC m=+86.604829752 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.354982 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:05.854962916 +0000 UTC m=+86.604877643 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.355282 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.355684 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.356361 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.356555 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.356826 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.356973 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.357205 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.357390 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.358046 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.359564 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.359788 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.359816 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.360159 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.367122 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.367559 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.367638 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.371487 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.372814 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.387590 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.392978 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.423678 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.423916 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.424042 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.424131 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.424363 4930 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.424545 4930 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.424658 4930 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.424752 4930 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.424844 4930 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.424933 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.425037 4930 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.425134 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.425217 4930 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.425302 4930 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.425403 4930 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.425527 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.425637 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.425719 4930 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.425792 4930 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.425921 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.426009 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.426128 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.426252 4930 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.426485 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.426579 4930 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.426665 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.426744 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.426818 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.426895 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.426967 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.427051 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.427129 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.427208 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.427285 4930 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.427368 4930 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.427471 4930 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.427614 4930 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.427703 4930 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.427800 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.427899 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.427977 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.428059 4930 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.428134 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.428211 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.428289 4930 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.428369 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.428468 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.428554 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.428623 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.428693 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.428774 4930 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.428852 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.428931 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429015 4930 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429092 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429161 4930 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429245 4930 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429320 4930 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429391 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429499 4930 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429581 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429650 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429731 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429802 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429898 4930 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.429979 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430061 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430138 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430210 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430279 4930 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430348 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430423 4930 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430532 4930 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430620 4930 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430698 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430778 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430852 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430921 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.430989 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431064 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431142 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431211 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431284 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431355 4930 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431423 4930 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431539 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431616 4930 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431692 4930 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431765 4930 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431836 4930 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431918 4930 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.431989 4930 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.432063 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.432136 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.432207 4930 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.432278 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.432353 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.432460 4930 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.432580 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.432683 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.432761 4930 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.432844 4930 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.432922 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.433063 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.433160 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.433245 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.433328 4930 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.433421 4930 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.433560 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.433709 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.433798 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.433894 4930 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.433968 4930 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.434054 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.434149 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.434232 4930 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.434314 4930 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.434392 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.434517 4930 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.434609 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.434689 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.434762 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.434845 4930 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.434932 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.435085 4930 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.435223 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.435363 4930 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.435534 4930 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.435701 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.435847 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.436071 4930 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.436251 4930 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.436468 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.436610 4930 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.436688 4930 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.436856 4930 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.436930 4930 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.437001 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.437080 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.437169 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.437273 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.437383 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.437510 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.437598 4930 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.437670 4930 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.437738 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.437853 4930 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.437941 4930 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.438039 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.438117 4930 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.438186 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.438256 4930 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.455513 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.455743 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.455833 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.455917 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.456000 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:05Z","lastTransitionTime":"2026-03-13T09:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.525022 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.539343 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.549754 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 13 09:14:05 crc kubenswrapper[4930]: W0313 09:14:05.555607 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-4b21294b3b3c56d608b302df1f8ca385f03fcbdeef5f72c16abe3a9055000a89 WatchSource:0}: Error finding container 4b21294b3b3c56d608b302df1f8ca385f03fcbdeef5f72c16abe3a9055000a89: Status 404 returned error can't find the container with id 4b21294b3b3c56d608b302df1f8ca385f03fcbdeef5f72c16abe3a9055000a89 Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.557678 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.557708 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.557716 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.557730 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.557740 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:05Z","lastTransitionTime":"2026-03-13T09:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:05 crc kubenswrapper[4930]: W0313 09:14:05.570544 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-b1f00272a3b13042f79646749df35a4f64176e0313ab6364db16aa835b190f68 WatchSource:0}: Error finding container b1f00272a3b13042f79646749df35a4f64176e0313ab6364db16aa835b190f68: Status 404 returned error can't find the container with id b1f00272a3b13042f79646749df35a4f64176e0313ab6364db16aa835b190f68 Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.661879 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.662239 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.662249 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.662263 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.662272 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:05Z","lastTransitionTime":"2026-03-13T09:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.764905 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.764934 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.764942 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.764955 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.764964 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:05Z","lastTransitionTime":"2026-03-13T09:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.840175 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.840303 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:06.840276804 +0000 UTC m=+87.590191491 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.840389 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.840515 4930 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.840562 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:06.840553481 +0000 UTC m=+87.590468168 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.867394 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.867444 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.867455 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.867470 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.867478 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:05Z","lastTransitionTime":"2026-03-13T09:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.941240 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.941308 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.941356 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.941538 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.941565 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.941582 4930 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.941646 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:06.941626385 +0000 UTC m=+87.691541072 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.942055 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.942122 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.942138 4930 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.942175 4930 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.942191 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:06.942175849 +0000 UTC m=+87.692090536 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:05 crc kubenswrapper[4930]: E0313 09:14:05.942279 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:06.942257211 +0000 UTC m=+87.692171958 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.970253 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.970313 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.970330 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.970349 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.970361 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:05Z","lastTransitionTime":"2026-03-13T09:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.975062 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.975886 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.977663 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.978877 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.980215 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.981752 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.982763 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.983659 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.985207 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.986231 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.987896 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.989176 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.990746 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.991599 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.992913 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.993867 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.995015 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.996648 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.997904 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 13 09:14:05 crc kubenswrapper[4930]: I0313 09:14:05.998887 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.000215 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.001812 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.003494 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.005016 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.005830 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.007687 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.009557 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.010634 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.011757 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.013315 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.014423 4930 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.014868 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.020250 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.021328 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.023177 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.026578 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.027653 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.028422 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.029898 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.030653 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.031571 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.032186 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.033389 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.034698 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.035317 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.036217 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.036817 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.038066 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.038595 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.039107 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.040130 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.040734 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.041726 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.042264 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.072404 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.072464 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.072477 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.072497 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.072509 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:06Z","lastTransitionTime":"2026-03-13T09:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.174966 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.175014 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.175025 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.175041 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.175051 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:06Z","lastTransitionTime":"2026-03-13T09:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.276948 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.277231 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.277313 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.277380 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.277457 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:06Z","lastTransitionTime":"2026-03-13T09:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.337460 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"64051a98b17412cc4bb68b87b37fa23427261ff1f22531ac0c0e69a098409022"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.337525 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"69bb65ece34b5c45b596df1707b4af90e5c35737cd02cee850c3b5c93fea664e"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.338194 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"b1f00272a3b13042f79646749df35a4f64176e0313ab6364db16aa835b190f68"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.339699 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"0f53b7e3a21874c64e257db8da16ff16c92321478719d748b8e7d2ffa8cc4c63"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.339777 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"90fc052b2b87e41422243f373943c80496a01a47ae6e35cc255898d1f89e03d7"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.339848 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"4b21294b3b3c56d608b302df1f8ca385f03fcbdeef5f72c16abe3a9055000a89"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.357301 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64051a98b17412cc4bb68b87b37fa23427261ff1f22531ac0c0e69a098409022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T09:14:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.374914 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.380477 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.380581 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.380658 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.380739 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.380868 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:06Z","lastTransitionTime":"2026-03-13T09:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.391423 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.407796 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.422886 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.437943 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.451612 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8dee6b1-215b-445a-9580-c3858a84e940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1e3de75ed03b7e44185eb7ab8df78138aeb01abc9c764309175f6df0433ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T09:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41d6b9b5f204b817b33c875a43ee086935dbe7ee9a2186711c964e664dd3747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b41d6b9b5f204b817b33c875a43ee086935dbe7ee9a2186711c964e664dd3747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T09:12:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T09:12:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T09:12:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.462534 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8dee6b1-215b-445a-9580-c3858a84e940\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-13T09:12:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bb1e3de75ed03b7e44185eb7ab8df78138aeb01abc9c764309175f6df0433ade\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T09:12:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b41d6b9b5f204b817b33c875a43ee086935dbe7ee9a2186711c964e664dd3747\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b41d6b9b5f204b817b33c875a43ee086935dbe7ee9a2186711c964e664dd3747\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-13T09:12:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-13T09:12:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-13T09:12:40Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.479010 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.482362 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.482483 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.482501 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.482526 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.482544 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:06Z","lastTransitionTime":"2026-03-13T09:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.494826 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0f53b7e3a21874c64e257db8da16ff16c92321478719d748b8e7d2ffa8cc4c63\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T09:14:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://90fc052b2b87e41422243f373943c80496a01a47ae6e35cc255898d1f89e03d7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T09:14:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.511050 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://64051a98b17412cc4bb68b87b37fa23427261ff1f22531ac0c0e69a098409022\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-13T09:14:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.522863 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.538213 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.551162 4930 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-13T09:14:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-13T09:14:06Z is after 2025-08-24T17:21:41Z" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.585269 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.585619 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.585762 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.585893 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.586029 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:06Z","lastTransitionTime":"2026-03-13T09:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.688611 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.688661 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.688673 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.688691 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.688704 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:06Z","lastTransitionTime":"2026-03-13T09:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.791113 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.791148 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.791156 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.791170 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.791179 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:06Z","lastTransitionTime":"2026-03-13T09:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.848185 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.848291 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.848377 4930 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.848381 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:08.848351992 +0000 UTC m=+89.598266669 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.848457 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:08.848422484 +0000 UTC m=+89.598337211 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.896473 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.896707 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.896772 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.896843 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.896920 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:06Z","lastTransitionTime":"2026-03-13T09:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.949480 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.949532 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.949572 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.949709 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.949730 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.949743 4930 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.949793 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:08.949777955 +0000 UTC m=+89.699692642 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.949797 4930 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.949864 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:08.949846017 +0000 UTC m=+89.699760694 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.949923 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.949933 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.949943 4930 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.949963 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:08.94995733 +0000 UTC m=+89.699872007 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.971229 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.971300 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.971536 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.971598 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.971766 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.971979 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.979830 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.979931 4930 scope.go:117] "RemoveContainer" containerID="1fc8d89dc87d04d45f03cdf66b778aceb7d3a920f5fd254edc0f7d5491c145bf" Mar 13 09:14:06 crc kubenswrapper[4930]: E0313 09:14:06.980091 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.999146 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.999191 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.999202 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.999219 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:06 crc kubenswrapper[4930]: I0313 09:14:06.999231 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:06Z","lastTransitionTime":"2026-03-13T09:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.101677 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.101715 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.101723 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.101737 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.101745 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:07Z","lastTransitionTime":"2026-03-13T09:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.203650 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.203683 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.203691 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.203703 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.203711 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:07Z","lastTransitionTime":"2026-03-13T09:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.305368 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.305621 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.305692 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.305769 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.305855 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:07Z","lastTransitionTime":"2026-03-13T09:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.342974 4930 scope.go:117] "RemoveContainer" containerID="1fc8d89dc87d04d45f03cdf66b778aceb7d3a920f5fd254edc0f7d5491c145bf" Mar 13 09:14:07 crc kubenswrapper[4930]: E0313 09:14:07.343282 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.408367 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.408405 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.408416 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.408449 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.408460 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:07Z","lastTransitionTime":"2026-03-13T09:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.510380 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.510421 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.510452 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.510471 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.510485 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:07Z","lastTransitionTime":"2026-03-13T09:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.613496 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.613540 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.613551 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.613569 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.613582 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:07Z","lastTransitionTime":"2026-03-13T09:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.715876 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.715938 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.715958 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.715982 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.716000 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:07Z","lastTransitionTime":"2026-03-13T09:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.818768 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.818831 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.818843 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.818862 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.818874 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:07Z","lastTransitionTime":"2026-03-13T09:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.920852 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.920901 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.920915 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.920935 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:07 crc kubenswrapper[4930]: I0313 09:14:07.920949 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:07Z","lastTransitionTime":"2026-03-13T09:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.023353 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.023416 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.023462 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.023487 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.023504 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:08Z","lastTransitionTime":"2026-03-13T09:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.126114 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.126171 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.126187 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.126216 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.126234 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:08Z","lastTransitionTime":"2026-03-13T09:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.229010 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.229053 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.229073 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.229097 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.229114 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:08Z","lastTransitionTime":"2026-03-13T09:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.332134 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.332176 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.332185 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.332204 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.332216 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:08Z","lastTransitionTime":"2026-03-13T09:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.345711 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8e88045243df441f45b0cf8aa8475600473e9cefeb05bb8af537decf70336279"} Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.418333 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=4.418307066 podStartE2EDuration="4.418307066s" podCreationTimestamp="2026-03-13 09:14:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:08.418194303 +0000 UTC m=+89.168109000" watchObservedRunningTime="2026-03-13 09:14:08.418307066 +0000 UTC m=+89.168221773" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.434494 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.434552 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.434562 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.434575 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.434584 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:08Z","lastTransitionTime":"2026-03-13T09:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.537383 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.537446 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.537460 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.537473 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.537482 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:08Z","lastTransitionTime":"2026-03-13T09:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.640325 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.640386 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.640406 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.640429 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.640509 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:08Z","lastTransitionTime":"2026-03-13T09:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.743183 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.743240 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.743257 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.743280 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.743298 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:08Z","lastTransitionTime":"2026-03-13T09:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.845790 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.845836 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.845852 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.845874 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.845891 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:08Z","lastTransitionTime":"2026-03-13T09:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.868001 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.868113 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.868236 4930 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.868303 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:12.868282066 +0000 UTC m=+93.618196773 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.868779 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:12.868761169 +0000 UTC m=+93.618675876 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.948561 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.948595 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.948606 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.948621 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.948633 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:08Z","lastTransitionTime":"2026-03-13T09:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.968567 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.968613 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.968642 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.968722 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.968734 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.968760 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.968764 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.968778 4930 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.968778 4930 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.968815 4930 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.968838 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:12.968819756 +0000 UTC m=+93.718734433 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.968947 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:12.968918809 +0000 UTC m=+93.718833516 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.968978 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:12.9689638 +0000 UTC m=+93.718878517 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.969860 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.969971 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.969990 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:08 crc kubenswrapper[4930]: I0313 09:14:08.970038 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.970067 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 09:14:08 crc kubenswrapper[4930]: E0313 09:14:08.970167 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.051353 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.051399 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.051412 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.051447 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.051460 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:09Z","lastTransitionTime":"2026-03-13T09:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.153981 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.154049 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.154066 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.154089 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.154106 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:09Z","lastTransitionTime":"2026-03-13T09:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.257600 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.257665 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.257679 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.257694 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.257707 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:09Z","lastTransitionTime":"2026-03-13T09:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.360253 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.360298 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.360311 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.360331 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.360343 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:09Z","lastTransitionTime":"2026-03-13T09:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.463020 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.463085 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.463104 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.463132 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.463152 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:09Z","lastTransitionTime":"2026-03-13T09:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.565815 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.565876 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.565902 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.565931 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.565952 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:09Z","lastTransitionTime":"2026-03-13T09:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.667913 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.668002 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.668020 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.668081 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.668100 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:09Z","lastTransitionTime":"2026-03-13T09:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.770603 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.770647 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.770659 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.770674 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.770683 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:09Z","lastTransitionTime":"2026-03-13T09:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.873227 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.873260 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.873268 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.873280 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.873289 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:09Z","lastTransitionTime":"2026-03-13T09:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.976067 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.976128 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.976145 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.976166 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:09 crc kubenswrapper[4930]: I0313 09:14:09.976183 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:09Z","lastTransitionTime":"2026-03-13T09:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.077501 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.077541 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.077549 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.077563 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.077572 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:10Z","lastTransitionTime":"2026-03-13T09:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.179651 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.179682 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.179692 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.179707 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.179718 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:10Z","lastTransitionTime":"2026-03-13T09:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.186448 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.186476 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.186483 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.186493 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.186502 4930 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-13T09:14:10Z","lastTransitionTime":"2026-03-13T09:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.970419 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.970498 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:10 crc kubenswrapper[4930]: I0313 09:14:10.970513 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:10 crc kubenswrapper[4930]: E0313 09:14:10.970567 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 09:14:10 crc kubenswrapper[4930]: E0313 09:14:10.970693 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 09:14:10 crc kubenswrapper[4930]: E0313 09:14:10.970773 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.201503 4930 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.208592 4930 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.966622 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-fv6sw"] Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.967064 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-fv6sw" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.969609 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.969689 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.970192 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.988529 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-724mj"] Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.988956 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-b2g6h"] Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.989112 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.989371 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-b2g6h" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.989495 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-br7r2"] Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.990499 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.992263 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.992495 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.992845 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.994778 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.995022 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.995242 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.995426 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.995882 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.996157 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.996343 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.996944 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 13 09:14:11 crc kubenswrapper[4930]: I0313 09:14:11.998581 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.005087 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nvxdn"] Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.006523 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.007704 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.007886 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.008148 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.008642 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.008971 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.009166 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.011740 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.079873 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k"] Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.080192 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.082605 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.082669 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.083552 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.085805 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098506 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-etc-openvswitch\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098541 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a92650e8-d2ad-4d63-a21b-998c19841660-multus-daemon-config\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098555 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-systemd-units\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098571 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlz24\" (UniqueName: \"kubernetes.io/projected/82ea51ef-4b24-4137-a011-27b26f759090-kube-api-access-vlz24\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098591 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-multus-cni-dir\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098605 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-var-lib-cni-bin\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098618 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-var-lib-kubelet\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098631 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zfld\" (UniqueName: \"kubernetes.io/projected/a92650e8-d2ad-4d63-a21b-998c19841660-kube-api-access-7zfld\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098646 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-node-log\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098663 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtcxv\" (UniqueName: \"kubernetes.io/projected/6bd55895-a75e-46de-bf2d-ed21c33217ea-kube-api-access-rtcxv\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098678 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22188dce-43d2-4c7e-aa9b-7090a71eeb06-mcd-auth-proxy-config\") pod \"machine-config-daemon-724mj\" (UID: \"22188dce-43d2-4c7e-aa9b-7090a71eeb06\") " pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098690 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-var-lib-cni-multus\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098703 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-etc-kubernetes\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098716 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098733 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-run-netns\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098747 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-systemd\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098762 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-var-lib-openvswitch\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098775 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-cni-netd\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098796 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-cnibin\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098811 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/82ea51ef-4b24-4137-a011-27b26f759090-os-release\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098824 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/82ea51ef-4b24-4137-a011-27b26f759090-tuning-conf-dir\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098838 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-slash\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098852 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovnkube-script-lib\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098867 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/22188dce-43d2-4c7e-aa9b-7090a71eeb06-proxy-tls\") pod \"machine-config-daemon-724mj\" (UID: \"22188dce-43d2-4c7e-aa9b-7090a71eeb06\") " pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098880 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-os-release\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098893 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-run-k8s-cni-cncf-io\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098908 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/82ea51ef-4b24-4137-a011-27b26f759090-cnibin\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098920 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-ovn\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098935 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovn-node-metrics-cert\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098948 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-system-cni-dir\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098960 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a92650e8-d2ad-4d63-a21b-998c19841660-cni-binary-copy\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098974 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/82ea51ef-4b24-4137-a011-27b26f759090-cni-binary-copy\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.098987 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/22188dce-43d2-4c7e-aa9b-7090a71eeb06-rootfs\") pod \"machine-config-daemon-724mj\" (UID: \"22188dce-43d2-4c7e-aa9b-7090a71eeb06\") " pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099006 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-run-netns\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099041 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-multus-conf-dir\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099057 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-env-overrides\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099076 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44vm6\" (UniqueName: \"kubernetes.io/projected/d67a1e7c-4f81-4c2c-aaa9-03189438b7cd-kube-api-access-44vm6\") pod \"node-resolver-fv6sw\" (UID: \"d67a1e7c-4f81-4c2c-aaa9-03189438b7cd\") " pod="openshift-dns/node-resolver-fv6sw" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099089 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-cni-bin\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099104 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-multus-socket-dir-parent\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099118 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovnkube-config\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099139 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-log-socket\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099153 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-run-ovn-kubernetes\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099172 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-kubelet\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099186 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-openvswitch\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099200 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-hostroot\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099213 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/82ea51ef-4b24-4137-a011-27b26f759090-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099228 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkpjx\" (UniqueName: \"kubernetes.io/projected/22188dce-43d2-4c7e-aa9b-7090a71eeb06-kube-api-access-kkpjx\") pod \"machine-config-daemon-724mj\" (UID: \"22188dce-43d2-4c7e-aa9b-7090a71eeb06\") " pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099242 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-run-multus-certs\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099256 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/82ea51ef-4b24-4137-a011-27b26f759090-system-cni-dir\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.099271 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d67a1e7c-4f81-4c2c-aaa9-03189438b7cd-hosts-file\") pod \"node-resolver-fv6sw\" (UID: \"d67a1e7c-4f81-4c2c-aaa9-03189438b7cd\") " pod="openshift-dns/node-resolver-fv6sw" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.198060 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-jm2fs"] Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.198406 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jm2fs" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.199861 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a92650e8-d2ad-4d63-a21b-998c19841660-multus-daemon-config\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.199894 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-systemd-units\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.199918 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-etc-openvswitch\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.199943 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-multus-cni-dir\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.199967 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-var-lib-cni-bin\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.199992 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlz24\" (UniqueName: \"kubernetes.io/projected/82ea51ef-4b24-4137-a011-27b26f759090-kube-api-access-vlz24\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200015 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zfld\" (UniqueName: \"kubernetes.io/projected/a92650e8-d2ad-4d63-a21b-998c19841660-kube-api-access-7zfld\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200038 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-node-log\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200060 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtcxv\" (UniqueName: \"kubernetes.io/projected/6bd55895-a75e-46de-bf2d-ed21c33217ea-kube-api-access-rtcxv\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200061 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-var-lib-cni-bin\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200085 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22188dce-43d2-4c7e-aa9b-7090a71eeb06-mcd-auth-proxy-config\") pod \"machine-config-daemon-724mj\" (UID: \"22188dce-43d2-4c7e-aa9b-7090a71eeb06\") " pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200110 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-var-lib-kubelet\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200114 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-node-log\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200132 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-etc-kubernetes\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200162 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200183 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-var-lib-cni-multus\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200188 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-var-lib-kubelet\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200199 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-systemd\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200061 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-etc-openvswitch\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200214 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-var-lib-openvswitch\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200227 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-cni-netd\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200233 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200231 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-multus-cni-dir\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200265 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-var-lib-cni-multus\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200277 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-cnibin\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200270 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-var-lib-openvswitch\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200293 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-systemd\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200249 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-cnibin\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200349 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/82ea51ef-4b24-4137-a011-27b26f759090-os-release\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200366 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-run-netns\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200382 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-slash\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200302 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-cni-netd\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200401 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovnkube-script-lib\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200420 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/82ea51ef-4b24-4137-a011-27b26f759090-os-release\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200423 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/61c975d1-f6b7-49d5-9349-5d14a338bc64-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200445 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-run-netns\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200476 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/22188dce-43d2-4c7e-aa9b-7090a71eeb06-proxy-tls\") pod \"machine-config-daemon-724mj\" (UID: \"22188dce-43d2-4c7e-aa9b-7090a71eeb06\") " pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200495 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-os-release\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200512 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-run-k8s-cni-cncf-io\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200528 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/82ea51ef-4b24-4137-a011-27b26f759090-tuning-conf-dir\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200545 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/82ea51ef-4b24-4137-a011-27b26f759090-cnibin\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200561 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovn-node-metrics-cert\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200578 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-system-cni-dir\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200592 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a92650e8-d2ad-4d63-a21b-998c19841660-cni-binary-copy\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200606 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/82ea51ef-4b24-4137-a011-27b26f759090-cni-binary-copy\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200619 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-ovn\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200630 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/a92650e8-d2ad-4d63-a21b-998c19841660-multus-daemon-config\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200656 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/22188dce-43d2-4c7e-aa9b-7090a71eeb06-rootfs\") pod \"machine-config-daemon-724mj\" (UID: \"22188dce-43d2-4c7e-aa9b-7090a71eeb06\") " pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200632 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/22188dce-43d2-4c7e-aa9b-7090a71eeb06-rootfs\") pod \"machine-config-daemon-724mj\" (UID: \"22188dce-43d2-4c7e-aa9b-7090a71eeb06\") " pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200707 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-os-release\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200711 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-multus-conf-dir\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200731 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-env-overrides\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200754 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44vm6\" (UniqueName: \"kubernetes.io/projected/d67a1e7c-4f81-4c2c-aaa9-03189438b7cd-kube-api-access-44vm6\") pod \"node-resolver-fv6sw\" (UID: \"d67a1e7c-4f81-4c2c-aaa9-03189438b7cd\") " pod="openshift-dns/node-resolver-fv6sw" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200766 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22188dce-43d2-4c7e-aa9b-7090a71eeb06-mcd-auth-proxy-config\") pod \"machine-config-daemon-724mj\" (UID: \"22188dce-43d2-4c7e-aa9b-7090a71eeb06\") " pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200476 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-slash\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200802 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-run-netns\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200165 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-etc-kubernetes\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200814 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-multus-conf-dir\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200772 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-run-netns\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200731 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-run-k8s-cni-cncf-io\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.200842 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/82ea51ef-4b24-4137-a011-27b26f759090-cnibin\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.201593 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/82ea51ef-4b24-4137-a011-27b26f759090-tuning-conf-dir\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.201642 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-systemd-units\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.201765 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovnkube-script-lib\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.202210 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.202262 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.202314 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.205654 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-ovn\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.205769 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-system-cni-dir\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.205804 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/22188dce-43d2-4c7e-aa9b-7090a71eeb06-proxy-tls\") pod \"machine-config-daemon-724mj\" (UID: \"22188dce-43d2-4c7e-aa9b-7090a71eeb06\") " pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.205811 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-multus-socket-dir-parent\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.205898 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-cni-bin\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.205914 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-multus-socket-dir-parent\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.205935 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovnkube-config\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206134 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206270 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/82ea51ef-4b24-4137-a011-27b26f759090-cni-binary-copy\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206291 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61c975d1-f6b7-49d5-9349-5d14a338bc64-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206328 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-env-overrides\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206358 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-log-socket\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206419 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-run-ovn-kubernetes\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206452 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-cni-bin\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206458 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/a92650e8-d2ad-4d63-a21b-998c19841660-cni-binary-copy\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206472 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/61c975d1-f6b7-49d5-9349-5d14a338bc64-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206639 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-log-socket\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206673 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-run-ovn-kubernetes\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206680 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61c975d1-f6b7-49d5-9349-5d14a338bc64-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206723 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-openvswitch\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206761 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovnkube-config\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206769 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/61c975d1-f6b7-49d5-9349-5d14a338bc64-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206799 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-openvswitch\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206831 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-hostroot\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206865 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/82ea51ef-4b24-4137-a011-27b26f759090-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206887 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-hostroot\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206930 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-kubelet\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206951 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-run-multus-certs\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206985 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/82ea51ef-4b24-4137-a011-27b26f759090-system-cni-dir\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.206993 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-kubelet\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.207006 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/a92650e8-d2ad-4d63-a21b-998c19841660-host-run-multus-certs\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.207052 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d67a1e7c-4f81-4c2c-aaa9-03189438b7cd-hosts-file\") pod \"node-resolver-fv6sw\" (UID: \"d67a1e7c-4f81-4c2c-aaa9-03189438b7cd\") " pod="openshift-dns/node-resolver-fv6sw" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.207088 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkpjx\" (UniqueName: \"kubernetes.io/projected/22188dce-43d2-4c7e-aa9b-7090a71eeb06-kube-api-access-kkpjx\") pod \"machine-config-daemon-724mj\" (UID: \"22188dce-43d2-4c7e-aa9b-7090a71eeb06\") " pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.207094 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/82ea51ef-4b24-4137-a011-27b26f759090-system-cni-dir\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.207151 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/d67a1e7c-4f81-4c2c-aaa9-03189438b7cd-hosts-file\") pod \"node-resolver-fv6sw\" (UID: \"d67a1e7c-4f81-4c2c-aaa9-03189438b7cd\") " pod="openshift-dns/node-resolver-fv6sw" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.207779 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/82ea51ef-4b24-4137-a011-27b26f759090-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.208647 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovn-node-metrics-cert\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.222034 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44vm6\" (UniqueName: \"kubernetes.io/projected/d67a1e7c-4f81-4c2c-aaa9-03189438b7cd-kube-api-access-44vm6\") pod \"node-resolver-fv6sw\" (UID: \"d67a1e7c-4f81-4c2c-aaa9-03189438b7cd\") " pod="openshift-dns/node-resolver-fv6sw" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.224343 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlz24\" (UniqueName: \"kubernetes.io/projected/82ea51ef-4b24-4137-a011-27b26f759090-kube-api-access-vlz24\") pod \"multus-additional-cni-plugins-br7r2\" (UID: \"82ea51ef-4b24-4137-a011-27b26f759090\") " pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.224907 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkpjx\" (UniqueName: \"kubernetes.io/projected/22188dce-43d2-4c7e-aa9b-7090a71eeb06-kube-api-access-kkpjx\") pod \"machine-config-daemon-724mj\" (UID: \"22188dce-43d2-4c7e-aa9b-7090a71eeb06\") " pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.228891 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtcxv\" (UniqueName: \"kubernetes.io/projected/6bd55895-a75e-46de-bf2d-ed21c33217ea-kube-api-access-rtcxv\") pod \"ovnkube-node-nvxdn\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.232865 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zfld\" (UniqueName: \"kubernetes.io/projected/a92650e8-d2ad-4d63-a21b-998c19841660-kube-api-access-7zfld\") pod \"multus-b2g6h\" (UID: \"a92650e8-d2ad-4d63-a21b-998c19841660\") " pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.294577 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-fv6sw" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.306877 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.308053 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/42ccbcb7-413e-401f-a43c-2ad7956a86f4-host\") pod \"node-ca-jm2fs\" (UID: \"42ccbcb7-413e-401f-a43c-2ad7956a86f4\") " pod="openshift-image-registry/node-ca-jm2fs" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.308075 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/42ccbcb7-413e-401f-a43c-2ad7956a86f4-serviceca\") pod \"node-ca-jm2fs\" (UID: \"42ccbcb7-413e-401f-a43c-2ad7956a86f4\") " pod="openshift-image-registry/node-ca-jm2fs" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.308098 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgkh9\" (UniqueName: \"kubernetes.io/projected/42ccbcb7-413e-401f-a43c-2ad7956a86f4-kube-api-access-lgkh9\") pod \"node-ca-jm2fs\" (UID: \"42ccbcb7-413e-401f-a43c-2ad7956a86f4\") " pod="openshift-image-registry/node-ca-jm2fs" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.308115 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/61c975d1-f6b7-49d5-9349-5d14a338bc64-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.308149 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61c975d1-f6b7-49d5-9349-5d14a338bc64-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.308164 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/61c975d1-f6b7-49d5-9349-5d14a338bc64-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.308183 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61c975d1-f6b7-49d5-9349-5d14a338bc64-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.308199 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/61c975d1-f6b7-49d5-9349-5d14a338bc64-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.308575 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/61c975d1-f6b7-49d5-9349-5d14a338bc64-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.308601 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/61c975d1-f6b7-49d5-9349-5d14a338bc64-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.308848 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/61c975d1-f6b7-49d5-9349-5d14a338bc64-service-ca\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.312498 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/61c975d1-f6b7-49d5-9349-5d14a338bc64-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.317994 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-b2g6h" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.324729 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/61c975d1-f6b7-49d5-9349-5d14a338bc64-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-fxm4k\" (UID: \"61c975d1-f6b7-49d5-9349-5d14a338bc64\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.337316 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-br7r2" Mar 13 09:14:12 crc kubenswrapper[4930]: W0313 09:14:12.338910 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda92650e8_d2ad_4d63_a21b_998c19841660.slice/crio-de690d36bde3aa017f9200887414b6f949354ca1166595c3dc542871332acd56 WatchSource:0}: Error finding container de690d36bde3aa017f9200887414b6f949354ca1166595c3dc542871332acd56: Status 404 returned error can't find the container with id de690d36bde3aa017f9200887414b6f949354ca1166595c3dc542871332acd56 Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.343771 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8"] Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.344151 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.345622 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.346004 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.351288 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:12 crc kubenswrapper[4930]: W0313 09:14:12.363000 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod82ea51ef_4b24_4137_a011_27b26f759090.slice/crio-6528da2cc4bae071ba2855d7ecf2e7a3ffb544063c257e86c76504c9830040c8 WatchSource:0}: Error finding container 6528da2cc4bae071ba2855d7ecf2e7a3ffb544063c257e86c76504c9830040c8: Status 404 returned error can't find the container with id 6528da2cc4bae071ba2855d7ecf2e7a3ffb544063c257e86c76504c9830040c8 Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.364874 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b2g6h" event={"ID":"a92650e8-d2ad-4d63-a21b-998c19841660","Type":"ContainerStarted","Data":"de690d36bde3aa017f9200887414b6f949354ca1166595c3dc542871332acd56"} Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.365904 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-vfjf8"] Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.366302 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:12 crc kubenswrapper[4930]: E0313 09:14:12.366363 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vfjf8" podUID="8298eba9-72fc-4be8-9736-defaad400dcd" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.369974 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"ce744c84596c64e8d7a6767b31dabb683e022ec225bf0cdafae291c69b14c608"} Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.372124 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-fv6sw" event={"ID":"d67a1e7c-4f81-4c2c-aaa9-03189438b7cd","Type":"ContainerStarted","Data":"ae9070ba3fe71e9cbac3db95793f69cafafb43fa7b8d9b04c547a45d197cafb8"} Mar 13 09:14:12 crc kubenswrapper[4930]: W0313 09:14:12.379413 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6bd55895_a75e_46de_bf2d_ed21c33217ea.slice/crio-df3e4b2d0b57c5f3f55e05e5bbda888a14f37a93cb548fad257a9e530889af86 WatchSource:0}: Error finding container df3e4b2d0b57c5f3f55e05e5bbda888a14f37a93cb548fad257a9e530889af86: Status 404 returned error can't find the container with id df3e4b2d0b57c5f3f55e05e5bbda888a14f37a93cb548fad257a9e530889af86 Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.394840 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.409364 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0e6bd092-5b5e-460f-9ab3-7fa61a15549e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-dmzv8\" (UID: \"0e6bd092-5b5e-460f-9ab3-7fa61a15549e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.409418 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs\") pod \"network-metrics-daemon-vfjf8\" (UID: \"8298eba9-72fc-4be8-9736-defaad400dcd\") " pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.409603 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/42ccbcb7-413e-401f-a43c-2ad7956a86f4-host\") pod \"node-ca-jm2fs\" (UID: \"42ccbcb7-413e-401f-a43c-2ad7956a86f4\") " pod="openshift-image-registry/node-ca-jm2fs" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.409627 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0e6bd092-5b5e-460f-9ab3-7fa61a15549e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-dmzv8\" (UID: \"0e6bd092-5b5e-460f-9ab3-7fa61a15549e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.409644 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0e6bd092-5b5e-460f-9ab3-7fa61a15549e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-dmzv8\" (UID: \"0e6bd092-5b5e-460f-9ab3-7fa61a15549e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.409662 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/42ccbcb7-413e-401f-a43c-2ad7956a86f4-serviceca\") pod \"node-ca-jm2fs\" (UID: \"42ccbcb7-413e-401f-a43c-2ad7956a86f4\") " pod="openshift-image-registry/node-ca-jm2fs" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.409682 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgkh9\" (UniqueName: \"kubernetes.io/projected/42ccbcb7-413e-401f-a43c-2ad7956a86f4-kube-api-access-lgkh9\") pod \"node-ca-jm2fs\" (UID: \"42ccbcb7-413e-401f-a43c-2ad7956a86f4\") " pod="openshift-image-registry/node-ca-jm2fs" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.409697 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-978dg\" (UniqueName: \"kubernetes.io/projected/0e6bd092-5b5e-460f-9ab3-7fa61a15549e-kube-api-access-978dg\") pod \"ovnkube-control-plane-749d76644c-dmzv8\" (UID: \"0e6bd092-5b5e-460f-9ab3-7fa61a15549e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.409721 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gvlxd\" (UniqueName: \"kubernetes.io/projected/8298eba9-72fc-4be8-9736-defaad400dcd-kube-api-access-gvlxd\") pod \"network-metrics-daemon-vfjf8\" (UID: \"8298eba9-72fc-4be8-9736-defaad400dcd\") " pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.409780 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/42ccbcb7-413e-401f-a43c-2ad7956a86f4-host\") pod \"node-ca-jm2fs\" (UID: \"42ccbcb7-413e-401f-a43c-2ad7956a86f4\") " pod="openshift-image-registry/node-ca-jm2fs" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.410497 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/42ccbcb7-413e-401f-a43c-2ad7956a86f4-serviceca\") pod \"node-ca-jm2fs\" (UID: \"42ccbcb7-413e-401f-a43c-2ad7956a86f4\") " pod="openshift-image-registry/node-ca-jm2fs" Mar 13 09:14:12 crc kubenswrapper[4930]: W0313 09:14:12.422952 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61c975d1_f6b7_49d5_9349_5d14a338bc64.slice/crio-2f030dfcd038a8a3949e5dcca2525752c33e4172a5016313b91f6068edc6191d WatchSource:0}: Error finding container 2f030dfcd038a8a3949e5dcca2525752c33e4172a5016313b91f6068edc6191d: Status 404 returned error can't find the container with id 2f030dfcd038a8a3949e5dcca2525752c33e4172a5016313b91f6068edc6191d Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.425157 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgkh9\" (UniqueName: \"kubernetes.io/projected/42ccbcb7-413e-401f-a43c-2ad7956a86f4-kube-api-access-lgkh9\") pod \"node-ca-jm2fs\" (UID: \"42ccbcb7-413e-401f-a43c-2ad7956a86f4\") " pod="openshift-image-registry/node-ca-jm2fs" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.510283 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs\") pod \"network-metrics-daemon-vfjf8\" (UID: \"8298eba9-72fc-4be8-9736-defaad400dcd\") " pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.510332 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0e6bd092-5b5e-460f-9ab3-7fa61a15549e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-dmzv8\" (UID: \"0e6bd092-5b5e-460f-9ab3-7fa61a15549e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.510354 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0e6bd092-5b5e-460f-9ab3-7fa61a15549e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-dmzv8\" (UID: \"0e6bd092-5b5e-460f-9ab3-7fa61a15549e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.510389 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-978dg\" (UniqueName: \"kubernetes.io/projected/0e6bd092-5b5e-460f-9ab3-7fa61a15549e-kube-api-access-978dg\") pod \"ovnkube-control-plane-749d76644c-dmzv8\" (UID: \"0e6bd092-5b5e-460f-9ab3-7fa61a15549e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.510412 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gvlxd\" (UniqueName: \"kubernetes.io/projected/8298eba9-72fc-4be8-9736-defaad400dcd-kube-api-access-gvlxd\") pod \"network-metrics-daemon-vfjf8\" (UID: \"8298eba9-72fc-4be8-9736-defaad400dcd\") " pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.510461 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0e6bd092-5b5e-460f-9ab3-7fa61a15549e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-dmzv8\" (UID: \"0e6bd092-5b5e-460f-9ab3-7fa61a15549e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: E0313 09:14:12.510878 4930 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 09:14:12 crc kubenswrapper[4930]: E0313 09:14:12.510929 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs podName:8298eba9-72fc-4be8-9736-defaad400dcd nodeName:}" failed. No retries permitted until 2026-03-13 09:14:13.010913663 +0000 UTC m=+93.760828340 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs") pod "network-metrics-daemon-vfjf8" (UID: "8298eba9-72fc-4be8-9736-defaad400dcd") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.511032 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/0e6bd092-5b5e-460f-9ab3-7fa61a15549e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-dmzv8\" (UID: \"0e6bd092-5b5e-460f-9ab3-7fa61a15549e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.511155 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/0e6bd092-5b5e-460f-9ab3-7fa61a15549e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-dmzv8\" (UID: \"0e6bd092-5b5e-460f-9ab3-7fa61a15549e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.516482 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/0e6bd092-5b5e-460f-9ab3-7fa61a15549e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-dmzv8\" (UID: \"0e6bd092-5b5e-460f-9ab3-7fa61a15549e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.522802 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-jm2fs" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.532051 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-978dg\" (UniqueName: \"kubernetes.io/projected/0e6bd092-5b5e-460f-9ab3-7fa61a15549e-kube-api-access-978dg\") pod \"ovnkube-control-plane-749d76644c-dmzv8\" (UID: \"0e6bd092-5b5e-460f-9ab3-7fa61a15549e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.532476 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gvlxd\" (UniqueName: \"kubernetes.io/projected/8298eba9-72fc-4be8-9736-defaad400dcd-kube-api-access-gvlxd\") pod \"network-metrics-daemon-vfjf8\" (UID: \"8298eba9-72fc-4be8-9736-defaad400dcd\") " pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:12 crc kubenswrapper[4930]: W0313 09:14:12.549608 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod42ccbcb7_413e_401f_a43c_2ad7956a86f4.slice/crio-0d52e4d87da1835e48baa46bd4f85342b61021c174e658aa880aebeff889498f WatchSource:0}: Error finding container 0d52e4d87da1835e48baa46bd4f85342b61021c174e658aa880aebeff889498f: Status 404 returned error can't find the container with id 0d52e4d87da1835e48baa46bd4f85342b61021c174e658aa880aebeff889498f Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.655286 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" Mar 13 09:14:12 crc kubenswrapper[4930]: W0313 09:14:12.668718 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e6bd092_5b5e_460f_9ab3_7fa61a15549e.slice/crio-6fe16004c369a0357ad5b825a3f3418b62a3d856cd42803381f7ece3cae05513 WatchSource:0}: Error finding container 6fe16004c369a0357ad5b825a3f3418b62a3d856cd42803381f7ece3cae05513: Status 404 returned error can't find the container with id 6fe16004c369a0357ad5b825a3f3418b62a3d856cd42803381f7ece3cae05513 Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.913992 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:12 crc kubenswrapper[4930]: E0313 09:14:12.914171 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:20.91414021 +0000 UTC m=+101.664054897 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.914408 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:12 crc kubenswrapper[4930]: E0313 09:14:12.914501 4930 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 09:14:12 crc kubenswrapper[4930]: E0313 09:14:12.914557 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:20.914542791 +0000 UTC m=+101.664457468 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.970270 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.970340 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:12 crc kubenswrapper[4930]: E0313 09:14:12.970406 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 09:14:12 crc kubenswrapper[4930]: I0313 09:14:12.970497 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:12 crc kubenswrapper[4930]: E0313 09:14:12.970615 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 09:14:12 crc kubenswrapper[4930]: E0313 09:14:12.970690 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.014917 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.014970 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.015005 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.015045 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs\") pod \"network-metrics-daemon-vfjf8\" (UID: \"8298eba9-72fc-4be8-9736-defaad400dcd\") " pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.015111 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.015144 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.015159 4930 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.015221 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:21.015202624 +0000 UTC m=+101.765117331 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.015242 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.015260 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.015272 4930 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.015318 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:21.015302206 +0000 UTC m=+101.765216883 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.015367 4930 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.015398 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:21.015388978 +0000 UTC m=+101.765303655 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.015159 4930 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.015447 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs podName:8298eba9-72fc-4be8-9736-defaad400dcd nodeName:}" failed. No retries permitted until 2026-03-13 09:14:14.015420479 +0000 UTC m=+94.765335156 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs") pod "network-metrics-daemon-vfjf8" (UID: "8298eba9-72fc-4be8-9736-defaad400dcd") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.378065 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-fv6sw" event={"ID":"d67a1e7c-4f81-4c2c-aaa9-03189438b7cd","Type":"ContainerStarted","Data":"5a4cf0dc548e9d2ee534399e2b545b2cfc75d47c77d9ee9f74bcc7a618321e24"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.379878 4930 generic.go:334] "Generic (PLEG): container finished" podID="82ea51ef-4b24-4137-a011-27b26f759090" containerID="c06196152290c2458ca0e8d005336765ce3f28b2e244d75413983f1d39ff596e" exitCode=0 Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.379996 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-br7r2" event={"ID":"82ea51ef-4b24-4137-a011-27b26f759090","Type":"ContainerDied","Data":"c06196152290c2458ca0e8d005336765ce3f28b2e244d75413983f1d39ff596e"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.380085 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-br7r2" event={"ID":"82ea51ef-4b24-4137-a011-27b26f759090","Type":"ContainerStarted","Data":"6528da2cc4bae071ba2855d7ecf2e7a3ffb544063c257e86c76504c9830040c8"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.381835 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" event={"ID":"0e6bd092-5b5e-460f-9ab3-7fa61a15549e","Type":"ContainerStarted","Data":"802d6baac521e33672138081556594ccf067cfd409ab64f1aa506ecbea38c70d"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.381869 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" event={"ID":"0e6bd092-5b5e-460f-9ab3-7fa61a15549e","Type":"ContainerStarted","Data":"265126200dd9d148bf5d2dbe601abd2f6698ff4ac1f07991ee1bf79bfb498243"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.381884 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" event={"ID":"0e6bd092-5b5e-460f-9ab3-7fa61a15549e","Type":"ContainerStarted","Data":"6fe16004c369a0357ad5b825a3f3418b62a3d856cd42803381f7ece3cae05513"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.383149 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" event={"ID":"61c975d1-f6b7-49d5-9349-5d14a338bc64","Type":"ContainerStarted","Data":"1c2bd5c6ce05b6a13e1ccbf7762bb821f47221d597606946c70bbf7f35f23089"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.383187 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" event={"ID":"61c975d1-f6b7-49d5-9349-5d14a338bc64","Type":"ContainerStarted","Data":"2f030dfcd038a8a3949e5dcca2525752c33e4172a5016313b91f6068edc6191d"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.384964 4930 generic.go:334] "Generic (PLEG): container finished" podID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerID="1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486" exitCode=0 Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.385047 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerDied","Data":"1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.385072 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerStarted","Data":"df3e4b2d0b57c5f3f55e05e5bbda888a14f37a93cb548fad257a9e530889af86"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.386882 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"ecd39acc37017145f06e1e8b3fc9961665c16ed9b958afb515506bdcac52e1bb"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.386904 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"75a32c3a6a592ba94cf2f17752d6cd696cc0c651141c5d595391ac31dc2f4185"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.388071 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jm2fs" event={"ID":"42ccbcb7-413e-401f-a43c-2ad7956a86f4","Type":"ContainerStarted","Data":"6f69cf51947b73dd8820e2eba8bd3d6eca7bca87b602efebf544113ec419d80f"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.388091 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-jm2fs" event={"ID":"42ccbcb7-413e-401f-a43c-2ad7956a86f4","Type":"ContainerStarted","Data":"0d52e4d87da1835e48baa46bd4f85342b61021c174e658aa880aebeff889498f"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.389358 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b2g6h" event={"ID":"a92650e8-d2ad-4d63-a21b-998c19841660","Type":"ContainerStarted","Data":"867a5cfb15a92126f39e08a86bc404cb589b0461be5b317f95f3058e61b4f8f9"} Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.396189 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-fv6sw" podStartSLOduration=47.396168293 podStartE2EDuration="47.396168293s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:13.395338441 +0000 UTC m=+94.145253128" watchObservedRunningTime="2026-03-13 09:14:13.396168293 +0000 UTC m=+94.146083010" Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.435538 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-fxm4k" podStartSLOduration=47.435515554 podStartE2EDuration="47.435515554s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:13.43379115 +0000 UTC m=+94.183705877" watchObservedRunningTime="2026-03-13 09:14:13.435515554 +0000 UTC m=+94.185430271" Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.464653 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-dmzv8" podStartSLOduration=46.46463194 podStartE2EDuration="46.46463194s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:13.464594959 +0000 UTC m=+94.214509656" watchObservedRunningTime="2026-03-13 09:14:13.46463194 +0000 UTC m=+94.214546637" Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.500231 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podStartSLOduration=47.500212704 podStartE2EDuration="47.500212704s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:13.480052651 +0000 UTC m=+94.229967338" watchObservedRunningTime="2026-03-13 09:14:13.500212704 +0000 UTC m=+94.250127371" Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.500752 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-jm2fs" podStartSLOduration=47.500746978 podStartE2EDuration="47.500746978s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:13.496145288 +0000 UTC m=+94.246059955" watchObservedRunningTime="2026-03-13 09:14:13.500746978 +0000 UTC m=+94.250661655" Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.519010 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-b2g6h" podStartSLOduration=47.518990171 podStartE2EDuration="47.518990171s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:13.518963931 +0000 UTC m=+94.268878628" watchObservedRunningTime="2026-03-13 09:14:13.518990171 +0000 UTC m=+94.268904848" Mar 13 09:14:13 crc kubenswrapper[4930]: I0313 09:14:13.969977 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:13 crc kubenswrapper[4930]: E0313 09:14:13.970654 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vfjf8" podUID="8298eba9-72fc-4be8-9736-defaad400dcd" Mar 13 09:14:14 crc kubenswrapper[4930]: I0313 09:14:14.024588 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs\") pod \"network-metrics-daemon-vfjf8\" (UID: \"8298eba9-72fc-4be8-9736-defaad400dcd\") " pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:14 crc kubenswrapper[4930]: E0313 09:14:14.024681 4930 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 09:14:14 crc kubenswrapper[4930]: E0313 09:14:14.024716 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs podName:8298eba9-72fc-4be8-9736-defaad400dcd nodeName:}" failed. No retries permitted until 2026-03-13 09:14:16.024704149 +0000 UTC m=+96.774618826 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs") pod "network-metrics-daemon-vfjf8" (UID: "8298eba9-72fc-4be8-9736-defaad400dcd") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 09:14:14 crc kubenswrapper[4930]: I0313 09:14:14.395315 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerStarted","Data":"156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d"} Mar 13 09:14:14 crc kubenswrapper[4930]: I0313 09:14:14.395652 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerStarted","Data":"12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a"} Mar 13 09:14:14 crc kubenswrapper[4930]: I0313 09:14:14.395664 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerStarted","Data":"8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8"} Mar 13 09:14:14 crc kubenswrapper[4930]: I0313 09:14:14.395674 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerStarted","Data":"188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1"} Mar 13 09:14:14 crc kubenswrapper[4930]: I0313 09:14:14.395684 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerStarted","Data":"f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867"} Mar 13 09:14:14 crc kubenswrapper[4930]: I0313 09:14:14.395694 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerStarted","Data":"d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876"} Mar 13 09:14:14 crc kubenswrapper[4930]: I0313 09:14:14.397046 4930 generic.go:334] "Generic (PLEG): container finished" podID="82ea51ef-4b24-4137-a011-27b26f759090" containerID="b24d5b692c6403f446c6099aaf3e927f36dc392d35adeeddaaf86db0960a4e98" exitCode=0 Mar 13 09:14:14 crc kubenswrapper[4930]: I0313 09:14:14.397153 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-br7r2" event={"ID":"82ea51ef-4b24-4137-a011-27b26f759090","Type":"ContainerDied","Data":"b24d5b692c6403f446c6099aaf3e927f36dc392d35adeeddaaf86db0960a4e98"} Mar 13 09:14:14 crc kubenswrapper[4930]: I0313 09:14:14.970372 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:14 crc kubenswrapper[4930]: E0313 09:14:14.970548 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 09:14:14 crc kubenswrapper[4930]: I0313 09:14:14.971048 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:14 crc kubenswrapper[4930]: E0313 09:14:14.971132 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 09:14:14 crc kubenswrapper[4930]: I0313 09:14:14.971194 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:14 crc kubenswrapper[4930]: E0313 09:14:14.971263 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 09:14:15 crc kubenswrapper[4930]: I0313 09:14:15.401871 4930 generic.go:334] "Generic (PLEG): container finished" podID="82ea51ef-4b24-4137-a011-27b26f759090" containerID="782b3f7eaa1b4b41819591fdae7e95b106f6deea7fb83d2165ae590a0f5ffb08" exitCode=0 Mar 13 09:14:15 crc kubenswrapper[4930]: I0313 09:14:15.401933 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-br7r2" event={"ID":"82ea51ef-4b24-4137-a011-27b26f759090","Type":"ContainerDied","Data":"782b3f7eaa1b4b41819591fdae7e95b106f6deea7fb83d2165ae590a0f5ffb08"} Mar 13 09:14:15 crc kubenswrapper[4930]: I0313 09:14:15.970187 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:15 crc kubenswrapper[4930]: E0313 09:14:15.970576 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vfjf8" podUID="8298eba9-72fc-4be8-9736-defaad400dcd" Mar 13 09:14:16 crc kubenswrapper[4930]: I0313 09:14:16.043491 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs\") pod \"network-metrics-daemon-vfjf8\" (UID: \"8298eba9-72fc-4be8-9736-defaad400dcd\") " pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:16 crc kubenswrapper[4930]: E0313 09:14:16.043609 4930 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 09:14:16 crc kubenswrapper[4930]: E0313 09:14:16.043659 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs podName:8298eba9-72fc-4be8-9736-defaad400dcd nodeName:}" failed. No retries permitted until 2026-03-13 09:14:20.043645377 +0000 UTC m=+100.793560054 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs") pod "network-metrics-daemon-vfjf8" (UID: "8298eba9-72fc-4be8-9736-defaad400dcd") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 09:14:16 crc kubenswrapper[4930]: I0313 09:14:16.406967 4930 generic.go:334] "Generic (PLEG): container finished" podID="82ea51ef-4b24-4137-a011-27b26f759090" containerID="b53a3e9dc6045fd646ea70bc7ef93e7d42584bd8dfbe52c92b45ec38110429d1" exitCode=0 Mar 13 09:14:16 crc kubenswrapper[4930]: I0313 09:14:16.407008 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-br7r2" event={"ID":"82ea51ef-4b24-4137-a011-27b26f759090","Type":"ContainerDied","Data":"b53a3e9dc6045fd646ea70bc7ef93e7d42584bd8dfbe52c92b45ec38110429d1"} Mar 13 09:14:16 crc kubenswrapper[4930]: I0313 09:14:16.969867 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:16 crc kubenswrapper[4930]: I0313 09:14:16.969942 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:16 crc kubenswrapper[4930]: I0313 09:14:16.970285 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:16 crc kubenswrapper[4930]: E0313 09:14:16.970534 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 09:14:16 crc kubenswrapper[4930]: E0313 09:14:16.970453 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 09:14:16 crc kubenswrapper[4930]: E0313 09:14:16.970886 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 09:14:17 crc kubenswrapper[4930]: I0313 09:14:17.414231 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerStarted","Data":"a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049"} Mar 13 09:14:17 crc kubenswrapper[4930]: I0313 09:14:17.419456 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-br7r2" event={"ID":"82ea51ef-4b24-4137-a011-27b26f759090","Type":"ContainerDied","Data":"ad3f25465edd219d65514270252c33a9bb105fcb9ecdfc14faff6dfc911b4337"} Mar 13 09:14:17 crc kubenswrapper[4930]: I0313 09:14:17.419451 4930 generic.go:334] "Generic (PLEG): container finished" podID="82ea51ef-4b24-4137-a011-27b26f759090" containerID="ad3f25465edd219d65514270252c33a9bb105fcb9ecdfc14faff6dfc911b4337" exitCode=0 Mar 13 09:14:17 crc kubenswrapper[4930]: I0313 09:14:17.969831 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:17 crc kubenswrapper[4930]: E0313 09:14:17.969978 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vfjf8" podUID="8298eba9-72fc-4be8-9736-defaad400dcd" Mar 13 09:14:17 crc kubenswrapper[4930]: I0313 09:14:17.970559 4930 scope.go:117] "RemoveContainer" containerID="1fc8d89dc87d04d45f03cdf66b778aceb7d3a920f5fd254edc0f7d5491c145bf" Mar 13 09:14:17 crc kubenswrapper[4930]: E0313 09:14:17.970896 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 13 09:14:18 crc kubenswrapper[4930]: I0313 09:14:18.426276 4930 generic.go:334] "Generic (PLEG): container finished" podID="82ea51ef-4b24-4137-a011-27b26f759090" containerID="7f222c166a4a8d3986d8e5d3592a558ee35cfb657d2fdf662154cd39508f8449" exitCode=0 Mar 13 09:14:18 crc kubenswrapper[4930]: I0313 09:14:18.426331 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-br7r2" event={"ID":"82ea51ef-4b24-4137-a011-27b26f759090","Type":"ContainerDied","Data":"7f222c166a4a8d3986d8e5d3592a558ee35cfb657d2fdf662154cd39508f8449"} Mar 13 09:14:18 crc kubenswrapper[4930]: I0313 09:14:18.970001 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:18 crc kubenswrapper[4930]: I0313 09:14:18.970044 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:18 crc kubenswrapper[4930]: I0313 09:14:18.970081 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:18 crc kubenswrapper[4930]: E0313 09:14:18.970118 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 09:14:18 crc kubenswrapper[4930]: E0313 09:14:18.970424 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 09:14:18 crc kubenswrapper[4930]: E0313 09:14:18.970314 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 09:14:19 crc kubenswrapper[4930]: I0313 09:14:19.432863 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerStarted","Data":"af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594"} Mar 13 09:14:19 crc kubenswrapper[4930]: I0313 09:14:19.436248 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:19 crc kubenswrapper[4930]: I0313 09:14:19.436285 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:19 crc kubenswrapper[4930]: I0313 09:14:19.436308 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:19 crc kubenswrapper[4930]: I0313 09:14:19.441447 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-br7r2" event={"ID":"82ea51ef-4b24-4137-a011-27b26f759090","Type":"ContainerStarted","Data":"9cb2b3c47a61f3f9c9bc8a99d9d6b1c036be7c9846dd5ee139da21176282ca42"} Mar 13 09:14:19 crc kubenswrapper[4930]: I0313 09:14:19.462192 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:19 crc kubenswrapper[4930]: I0313 09:14:19.468325 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:19 crc kubenswrapper[4930]: I0313 09:14:19.503162 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" podStartSLOduration=53.50314478 podStartE2EDuration="53.50314478s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:19.471105318 +0000 UTC m=+100.221020035" watchObservedRunningTime="2026-03-13 09:14:19.50314478 +0000 UTC m=+100.253059457" Mar 13 09:14:19 crc kubenswrapper[4930]: I0313 09:14:19.523332 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-br7r2" podStartSLOduration=53.523315544 podStartE2EDuration="53.523315544s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:19.522085202 +0000 UTC m=+100.271999889" watchObservedRunningTime="2026-03-13 09:14:19.523315544 +0000 UTC m=+100.273230221" Mar 13 09:14:19 crc kubenswrapper[4930]: I0313 09:14:19.969898 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:19 crc kubenswrapper[4930]: E0313 09:14:19.971188 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vfjf8" podUID="8298eba9-72fc-4be8-9736-defaad400dcd" Mar 13 09:14:20 crc kubenswrapper[4930]: I0313 09:14:20.085941 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs\") pod \"network-metrics-daemon-vfjf8\" (UID: \"8298eba9-72fc-4be8-9736-defaad400dcd\") " pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:20 crc kubenswrapper[4930]: E0313 09:14:20.086066 4930 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 09:14:20 crc kubenswrapper[4930]: E0313 09:14:20.086116 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs podName:8298eba9-72fc-4be8-9736-defaad400dcd nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.086098653 +0000 UTC m=+108.836013330 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs") pod "network-metrics-daemon-vfjf8" (UID: "8298eba9-72fc-4be8-9736-defaad400dcd") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 13 09:14:20 crc kubenswrapper[4930]: I0313 09:14:20.966068 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vfjf8"] Mar 13 09:14:20 crc kubenswrapper[4930]: I0313 09:14:20.966241 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:20 crc kubenswrapper[4930]: E0313 09:14:20.966387 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vfjf8" podUID="8298eba9-72fc-4be8-9736-defaad400dcd" Mar 13 09:14:20 crc kubenswrapper[4930]: I0313 09:14:20.970403 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:20 crc kubenswrapper[4930]: I0313 09:14:20.970499 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:20 crc kubenswrapper[4930]: E0313 09:14:20.970566 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 09:14:20 crc kubenswrapper[4930]: E0313 09:14:20.970633 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 09:14:20 crc kubenswrapper[4930]: I0313 09:14:20.970684 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:20 crc kubenswrapper[4930]: E0313 09:14:20.970746 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 09:14:20 crc kubenswrapper[4930]: I0313 09:14:20.999032 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:20 crc kubenswrapper[4930]: I0313 09:14:20.999151 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:20 crc kubenswrapper[4930]: E0313 09:14:20.999273 4930 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 09:14:20 crc kubenswrapper[4930]: E0313 09:14:20.999329 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:36.999312639 +0000 UTC m=+117.749227326 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 13 09:14:20 crc kubenswrapper[4930]: E0313 09:14:20.999404 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:36.99935809 +0000 UTC m=+117.749272817 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:21 crc kubenswrapper[4930]: I0313 09:14:21.100147 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:21 crc kubenswrapper[4930]: I0313 09:14:21.100294 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:21 crc kubenswrapper[4930]: E0313 09:14:21.100321 4930 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 09:14:21 crc kubenswrapper[4930]: I0313 09:14:21.100342 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:21 crc kubenswrapper[4930]: E0313 09:14:21.100396 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:37.100379992 +0000 UTC m=+117.850294669 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 13 09:14:21 crc kubenswrapper[4930]: E0313 09:14:21.100561 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 09:14:21 crc kubenswrapper[4930]: E0313 09:14:21.100591 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 09:14:21 crc kubenswrapper[4930]: E0313 09:14:21.100612 4930 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:21 crc kubenswrapper[4930]: E0313 09:14:21.100697 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:37.10066669 +0000 UTC m=+117.850581407 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:21 crc kubenswrapper[4930]: E0313 09:14:21.100750 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 13 09:14:21 crc kubenswrapper[4930]: E0313 09:14:21.100763 4930 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 13 09:14:21 crc kubenswrapper[4930]: E0313 09:14:21.100773 4930 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:21 crc kubenswrapper[4930]: E0313 09:14:21.100797 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-13 09:14:37.100790363 +0000 UTC m=+117.850705040 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 13 09:14:22 crc kubenswrapper[4930]: I0313 09:14:22.970701 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:22 crc kubenswrapper[4930]: I0313 09:14:22.970744 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:22 crc kubenswrapper[4930]: I0313 09:14:22.970706 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:22 crc kubenswrapper[4930]: I0313 09:14:22.970812 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:22 crc kubenswrapper[4930]: E0313 09:14:22.970860 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 13 09:14:22 crc kubenswrapper[4930]: E0313 09:14:22.970906 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 13 09:14:22 crc kubenswrapper[4930]: E0313 09:14:22.970967 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 13 09:14:22 crc kubenswrapper[4930]: E0313 09:14:22.971134 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-vfjf8" podUID="8298eba9-72fc-4be8-9736-defaad400dcd" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.678743 4930 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.679107 4930 kubelet_node_status.go:538] "Fast updating node status as it just became ready" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.723861 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6sk6v"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.724475 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.726094 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xqlzn"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.726785 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.728886 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-9vpj8"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.729248 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.735420 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.735815 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.735845 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.735882 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.735970 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.736008 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.736150 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.737253 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-btdqm"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.737566 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.737647 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.738942 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.739382 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.740503 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nsh7c"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.740918 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.743145 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.744029 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.745588 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.746469 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.747307 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.747504 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.748317 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.749200 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f2kpv"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.750220 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.751976 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.752317 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.752494 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.756983 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.757583 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.760317 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.762605 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.766302 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.775904 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.776791 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.777115 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.777217 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.777397 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.777793 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.777879 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.778188 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.778398 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.778630 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.778649 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.778814 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.778855 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.779028 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.787216 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.787422 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.787595 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.787718 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.787767 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.787847 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.787987 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.788320 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.789273 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.789370 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.789530 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.789807 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.789865 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.789926 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.790001 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.790172 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.790567 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.790931 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.794569 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.794681 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.795139 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.795419 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.795675 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.795843 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796082 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796282 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796307 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796378 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796415 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796521 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796557 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796594 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796627 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796668 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796726 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796752 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796775 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796856 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796877 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796951 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796993 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.797034 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.797103 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.797270 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.797351 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.797390 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.797531 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.797656 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.797748 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.797779 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.797822 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.796525 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.797901 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.798014 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.798416 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-xnlmh"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.798963 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.806713 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.807048 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.808227 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.813267 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.813513 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.814272 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.814314 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.815813 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-gc8fp"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.816134 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.816309 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.816731 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.816793 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.836722 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.837268 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.837395 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.855483 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.856564 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.857126 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.858446 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.858907 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.859214 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-22ldf"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.859575 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.859618 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.859855 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860169 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860638 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860668 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngbzt\" (UniqueName: \"kubernetes.io/projected/1294227b-6b51-4269-93e2-542650c9dcb8-kube-api-access-ngbzt\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860689 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/afb89818-ec4f-4a94-8a7a-39e0e798616a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-d4ntr\" (UID: \"afb89818-ec4f-4a94-8a7a-39e0e798616a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860709 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b86a263-c28c-4b03-af1e-1e57f4f203b7-serving-cert\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860727 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjck8\" (UniqueName: \"kubernetes.io/projected/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-kube-api-access-fjck8\") pod \"route-controller-manager-6576b87f9c-c5xml\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860742 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/45e7c581-937a-46b2-9501-b704239b106c-node-pullsecrets\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860757 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-config\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860774 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1c70671b-0773-4c66-bc16-cf7a3e383572-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-z48bm\" (UID: \"1c70671b-0773-4c66-bc16-cf7a3e383572\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860789 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2668c\" (UniqueName: \"kubernetes.io/projected/0805c214-0e3a-44a0-8bfe-38e8adcccde4-kube-api-access-2668c\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860796 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860803 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860818 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-oauth-serving-cert\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860834 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15d7d52b-dc81-4400-bbce-353f00ff5103-serving-cert\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860849 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860863 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1294227b-6b51-4269-93e2-542650c9dcb8-etcd-client\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860877 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9srr9\" (UniqueName: \"kubernetes.io/projected/1c70671b-0773-4c66-bc16-cf7a3e383572-kube-api-access-9srr9\") pod \"cluster-image-registry-operator-dc59b4c8b-z48bm\" (UID: \"1c70671b-0773-4c66-bc16-cf7a3e383572\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860892 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f9ed5a87-d7c4-48bb-a331-afe0b8bc914a-images\") pod \"machine-api-operator-5694c8668f-6sk6v\" (UID: \"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860906 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860924 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.860963 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861022 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861047 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htl4k\" (UniqueName: \"kubernetes.io/projected/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-kube-api-access-htl4k\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861071 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c7a1eb51-28c1-46db-ab15-7d01d75e42e9-auth-proxy-config\") pod \"machine-approver-56656f9798-wmw8m\" (UID: \"c7a1eb51-28c1-46db-ab15-7d01d75e42e9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861096 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-service-ca\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861174 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861291 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861359 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861403 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861170 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1c70671b-0773-4c66-bc16-cf7a3e383572-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-z48bm\" (UID: \"1c70671b-0773-4c66-bc16-cf7a3e383572\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861481 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/45e7c581-937a-46b2-9501-b704239b106c-audit-dir\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861501 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-serving-cert\") pod \"route-controller-manager-6576b87f9c-c5xml\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861515 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qvsw\" (UniqueName: \"kubernetes.io/projected/f9ed5a87-d7c4-48bb-a331-afe0b8bc914a-kube-api-access-2qvsw\") pod \"machine-api-operator-5694c8668f-6sk6v\" (UID: \"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861540 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afb89818-ec4f-4a94-8a7a-39e0e798616a-serving-cert\") pod \"openshift-config-operator-7777fb866f-d4ntr\" (UID: \"afb89818-ec4f-4a94-8a7a-39e0e798616a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861554 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-oauth-config\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861569 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1294227b-6b51-4269-93e2-542650c9dcb8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861583 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-etcd-serving-ca\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861598 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15d7d52b-dc81-4400-bbce-353f00ff5103-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861614 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861628 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-audit\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861644 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-client-ca\") pod \"route-controller-manager-6576b87f9c-c5xml\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861658 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-image-import-ca\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861689 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj7cw\" (UniqueName: \"kubernetes.io/projected/afb89818-ec4f-4a94-8a7a-39e0e798616a-kube-api-access-nj7cw\") pod \"openshift-config-operator-7777fb866f-d4ntr\" (UID: \"afb89818-ec4f-4a94-8a7a-39e0e798616a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861711 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-config\") pod \"route-controller-manager-6576b87f9c-c5xml\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861727 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9ed5a87-d7c4-48bb-a331-afe0b8bc914a-config\") pod \"machine-api-operator-5694c8668f-6sk6v\" (UID: \"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861741 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15d7d52b-dc81-4400-bbce-353f00ff5103-service-ca-bundle\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861761 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861781 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz52n\" (UniqueName: \"kubernetes.io/projected/45e7c581-937a-46b2-9501-b704239b106c-kube-api-access-lz52n\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861805 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kttq\" (UniqueName: \"kubernetes.io/projected/15d7d52b-dc81-4400-bbce-353f00ff5103-kube-api-access-6kttq\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861822 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-trusted-ca-bundle\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861840 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861859 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/72724ed8-9cee-4f67-a055-c542e999b483-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-xl6pd\" (UID: \"72724ed8-9cee-4f67-a055-c542e999b483\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861876 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1294227b-6b51-4269-93e2-542650c9dcb8-serving-cert\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861891 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1294227b-6b51-4269-93e2-542650c9dcb8-audit-dir\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861908 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c7a1eb51-28c1-46db-ab15-7d01d75e42e9-machine-approver-tls\") pod \"machine-approver-56656f9798-wmw8m\" (UID: \"c7a1eb51-28c1-46db-ab15-7d01d75e42e9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861933 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861946 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-config\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861960 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1294227b-6b51-4269-93e2-542650c9dcb8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861973 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1294227b-6b51-4269-93e2-542650c9dcb8-encryption-config\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861987 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f9ed5a87-d7c4-48bb-a331-afe0b8bc914a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6sk6v\" (UID: \"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862000 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1294227b-6b51-4269-93e2-542650c9dcb8-audit-policies\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862014 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h7w2\" (UniqueName: \"kubernetes.io/projected/1b86a263-c28c-4b03-af1e-1e57f4f203b7-kube-api-access-9h7w2\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862029 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7a1eb51-28c1-46db-ab15-7d01d75e42e9-config\") pod \"machine-approver-56656f9798-wmw8m\" (UID: \"c7a1eb51-28c1-46db-ab15-7d01d75e42e9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862048 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jdpp\" (UniqueName: \"kubernetes.io/projected/72724ed8-9cee-4f67-a055-c542e999b483-kube-api-access-2jdpp\") pod \"cluster-samples-operator-665b6dd947-xl6pd\" (UID: \"72724ed8-9cee-4f67-a055-c542e999b483\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862061 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/45e7c581-937a-46b2-9501-b704239b106c-encryption-config\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862084 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-audit-dir\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862099 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-config\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862113 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-audit-policies\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.861294 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862127 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45e7c581-937a-46b2-9501-b704239b106c-serving-cert\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862140 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862155 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-client-ca\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862172 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15d7d52b-dc81-4400-bbce-353f00ff5103-config\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862187 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862205 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1c70671b-0773-4c66-bc16-cf7a3e383572-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-z48bm\" (UID: \"1c70671b-0773-4c66-bc16-cf7a3e383572\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862219 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862235 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-serving-cert\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862250 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/45e7c581-937a-46b2-9501-b704239b106c-etcd-client\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862263 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.862278 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jwj6f\" (UniqueName: \"kubernetes.io/projected/c7a1eb51-28c1-46db-ab15-7d01d75e42e9-kube-api-access-jwj6f\") pod \"machine-approver-56656f9798-wmw8m\" (UID: \"c7a1eb51-28c1-46db-ab15-7d01d75e42e9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.875549 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w957v"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.875959 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.876052 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-shpj8"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.876313 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.876444 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.876595 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.876636 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.876643 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.876816 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.876594 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.876938 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.877327 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.879501 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdbqj"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.879907 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.879976 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nptzn"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.880194 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.880705 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-nptzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.880920 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.881539 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.881709 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-kvzkk"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.882094 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kvzkk" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.882328 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.882782 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.883069 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.883712 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.885205 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.887172 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-smklb"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.887457 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xqlzn"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.887517 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-smklb" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.887678 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.887924 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.891166 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-l2c8z"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.891257 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.892495 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6sk6v"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.892907 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-l2c8z" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.893175 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.893917 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.894272 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.895382 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.895953 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.906810 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.909061 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.910986 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-9vpj8"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.913413 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.913970 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.914041 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.914294 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.917976 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-mpbsp"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.919228 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-496vm"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.919900 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.920172 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-mpbsp" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.920547 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f2kpv"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.923085 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.924947 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-rlvdx"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.925829 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rlvdx" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.926048 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-btdqm"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.927114 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.929909 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.930127 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.934069 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-xnlmh"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.936176 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.939398 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nptzn"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.940394 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w957v"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.942160 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.944714 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nsh7c"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.945944 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.946073 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-22ldf"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.946880 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-shpj8"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.947798 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.948863 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.949962 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.952003 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-w98w8"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.952651 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.952941 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2k7k2"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.954099 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdbqj"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.954246 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.954706 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.955677 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-kvzkk"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.956660 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-qnw8j"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.957102 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.957774 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-l2c8z"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.958614 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.959464 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-smklb"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.960326 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.961480 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.962561 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.962870 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.962895 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-config\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.962914 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f9ed5a87-d7c4-48bb-a331-afe0b8bc914a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6sk6v\" (UID: \"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.962931 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1294227b-6b51-4269-93e2-542650c9dcb8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.962947 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1294227b-6b51-4269-93e2-542650c9dcb8-encryption-config\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.962962 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7a1eb51-28c1-46db-ab15-7d01d75e42e9-config\") pod \"machine-approver-56656f9798-wmw8m\" (UID: \"c7a1eb51-28c1-46db-ab15-7d01d75e42e9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.962975 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1294227b-6b51-4269-93e2-542650c9dcb8-audit-policies\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.962998 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h7w2\" (UniqueName: \"kubernetes.io/projected/1b86a263-c28c-4b03-af1e-1e57f4f203b7-kube-api-access-9h7w2\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963017 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jdpp\" (UniqueName: \"kubernetes.io/projected/72724ed8-9cee-4f67-a055-c542e999b483-kube-api-access-2jdpp\") pod \"cluster-samples-operator-665b6dd947-xl6pd\" (UID: \"72724ed8-9cee-4f67-a055-c542e999b483\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963032 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/45e7c581-937a-46b2-9501-b704239b106c-encryption-config\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963058 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-audit-policies\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963073 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-audit-dir\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963089 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-config\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963103 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45e7c581-937a-46b2-9501-b704239b106c-serving-cert\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963121 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/805de6fb-8159-4c0d-bba1-f2d20558fb40-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nk9n9\" (UID: \"805de6fb-8159-4c0d-bba1-f2d20558fb40\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963136 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963152 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-client-ca\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963169 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15d7d52b-dc81-4400-bbce-353f00ff5103-config\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963184 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963204 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33b20f7f-12af-43bd-8937-426f36818e3d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-r9svm\" (UID: \"33b20f7f-12af-43bd-8937-426f36818e3d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963224 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1c70671b-0773-4c66-bc16-cf7a3e383572-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-z48bm\" (UID: \"1c70671b-0773-4c66-bc16-cf7a3e383572\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963242 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963261 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ntjsh\" (UniqueName: \"kubernetes.io/projected/33b20f7f-12af-43bd-8937-426f36818e3d-kube-api-access-ntjsh\") pod \"openshift-controller-manager-operator-756b6f6bc6-r9svm\" (UID: \"33b20f7f-12af-43bd-8937-426f36818e3d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963283 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-serving-cert\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963300 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/45e7c581-937a-46b2-9501-b704239b106c-etcd-client\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963315 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963332 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jwj6f\" (UniqueName: \"kubernetes.io/projected/c7a1eb51-28c1-46db-ab15-7d01d75e42e9-kube-api-access-jwj6f\") pod \"machine-approver-56656f9798-wmw8m\" (UID: \"c7a1eb51-28c1-46db-ab15-7d01d75e42e9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963347 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngbzt\" (UniqueName: \"kubernetes.io/projected/1294227b-6b51-4269-93e2-542650c9dcb8-kube-api-access-ngbzt\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963363 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/afb89818-ec4f-4a94-8a7a-39e0e798616a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-d4ntr\" (UID: \"afb89818-ec4f-4a94-8a7a-39e0e798616a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963378 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b86a263-c28c-4b03-af1e-1e57f4f203b7-serving-cert\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963395 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjck8\" (UniqueName: \"kubernetes.io/projected/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-kube-api-access-fjck8\") pod \"route-controller-manager-6576b87f9c-c5xml\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963410 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/45e7c581-937a-46b2-9501-b704239b106c-node-pullsecrets\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963425 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-config\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963456 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33b20f7f-12af-43bd-8937-426f36818e3d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-r9svm\" (UID: \"33b20f7f-12af-43bd-8937-426f36818e3d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963473 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1c70671b-0773-4c66-bc16-cf7a3e383572-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-z48bm\" (UID: \"1c70671b-0773-4c66-bc16-cf7a3e383572\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963475 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1294227b-6b51-4269-93e2-542650c9dcb8-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963487 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2668c\" (UniqueName: \"kubernetes.io/projected/0805c214-0e3a-44a0-8bfe-38e8adcccde4-kube-api-access-2668c\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963512 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15d7d52b-dc81-4400-bbce-353f00ff5103-serving-cert\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963535 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963556 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-oauth-serving-cert\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963604 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f920133-728b-43bb-a0ad-c851a2820b91-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dq2fw\" (UID: \"5f920133-728b-43bb-a0ad-c851a2820b91\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963632 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963654 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963675 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1294227b-6b51-4269-93e2-542650c9dcb8-etcd-client\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963696 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9srr9\" (UniqueName: \"kubernetes.io/projected/1c70671b-0773-4c66-bc16-cf7a3e383572-kube-api-access-9srr9\") pod \"cluster-image-registry-operator-dc59b4c8b-z48bm\" (UID: \"1c70671b-0773-4c66-bc16-cf7a3e383572\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963720 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f9ed5a87-d7c4-48bb-a331-afe0b8bc914a-images\") pod \"machine-api-operator-5694c8668f-6sk6v\" (UID: \"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963741 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1c70671b-0773-4c66-bc16-cf7a3e383572-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-z48bm\" (UID: \"1c70671b-0773-4c66-bc16-cf7a3e383572\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963744 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-config\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963762 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963787 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963812 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htl4k\" (UniqueName: \"kubernetes.io/projected/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-kube-api-access-htl4k\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963834 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c7a1eb51-28c1-46db-ab15-7d01d75e42e9-auth-proxy-config\") pod \"machine-approver-56656f9798-wmw8m\" (UID: \"c7a1eb51-28c1-46db-ab15-7d01d75e42e9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963854 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-service-ca\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963886 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/45e7c581-937a-46b2-9501-b704239b106c-audit-dir\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963909 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/805de6fb-8159-4c0d-bba1-f2d20558fb40-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nk9n9\" (UID: \"805de6fb-8159-4c0d-bba1-f2d20558fb40\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963943 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-serving-cert\") pod \"route-controller-manager-6576b87f9c-c5xml\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963968 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qvsw\" (UniqueName: \"kubernetes.io/projected/f9ed5a87-d7c4-48bb-a331-afe0b8bc914a-kube-api-access-2qvsw\") pod \"machine-api-operator-5694c8668f-6sk6v\" (UID: \"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963993 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afb89818-ec4f-4a94-8a7a-39e0e798616a-serving-cert\") pod \"openshift-config-operator-7777fb866f-d4ntr\" (UID: \"afb89818-ec4f-4a94-8a7a-39e0e798616a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964015 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-oauth-config\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964038 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1294227b-6b51-4269-93e2-542650c9dcb8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964061 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-etcd-serving-ca\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964084 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15d7d52b-dc81-4400-bbce-353f00ff5103-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964109 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f920133-728b-43bb-a0ad-c851a2820b91-config\") pod \"kube-apiserver-operator-766d6c64bb-dq2fw\" (UID: \"5f920133-728b-43bb-a0ad-c851a2820b91\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964135 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964156 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-audit\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964178 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-client-ca\") pod \"route-controller-manager-6576b87f9c-c5xml\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964203 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-image-import-ca\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964226 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj7cw\" (UniqueName: \"kubernetes.io/projected/afb89818-ec4f-4a94-8a7a-39e0e798616a-kube-api-access-nj7cw\") pod \"openshift-config-operator-7777fb866f-d4ntr\" (UID: \"afb89818-ec4f-4a94-8a7a-39e0e798616a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964247 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-config\") pod \"route-controller-manager-6576b87f9c-c5xml\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964267 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9ed5a87-d7c4-48bb-a331-afe0b8bc914a-config\") pod \"machine-api-operator-5694c8668f-6sk6v\" (UID: \"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964289 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15d7d52b-dc81-4400-bbce-353f00ff5103-service-ca-bundle\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964316 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jg9lc\" (UniqueName: \"kubernetes.io/projected/805de6fb-8159-4c0d-bba1-f2d20558fb40-kube-api-access-jg9lc\") pod \"kube-storage-version-migrator-operator-b67b599dd-nk9n9\" (UID: \"805de6fb-8159-4c0d-bba1-f2d20558fb40\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964341 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964364 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz52n\" (UniqueName: \"kubernetes.io/projected/45e7c581-937a-46b2-9501-b704239b106c-kube-api-access-lz52n\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964398 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kttq\" (UniqueName: \"kubernetes.io/projected/15d7d52b-dc81-4400-bbce-353f00ff5103-kube-api-access-6kttq\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964423 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-trusted-ca-bundle\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964464 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964488 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c7a1eb51-28c1-46db-ab15-7d01d75e42e9-machine-approver-tls\") pod \"machine-approver-56656f9798-wmw8m\" (UID: \"c7a1eb51-28c1-46db-ab15-7d01d75e42e9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964512 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/72724ed8-9cee-4f67-a055-c542e999b483-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-xl6pd\" (UID: \"72724ed8-9cee-4f67-a055-c542e999b483\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964534 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1294227b-6b51-4269-93e2-542650c9dcb8-serving-cert\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964555 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1294227b-6b51-4269-93e2-542650c9dcb8-audit-dir\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964579 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f920133-728b-43bb-a0ad-c851a2820b91-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-dq2fw\" (UID: \"5f920133-728b-43bb-a0ad-c851a2820b91\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964656 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/1c70671b-0773-4c66-bc16-cf7a3e383572-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-z48bm\" (UID: \"1c70671b-0773-4c66-bc16-cf7a3e383572\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.964981 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.965050 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-496vm"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.965565 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c7a1eb51-28c1-46db-ab15-7d01d75e42e9-auth-proxy-config\") pod \"machine-approver-56656f9798-wmw8m\" (UID: \"c7a1eb51-28c1-46db-ab15-7d01d75e42e9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.965601 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.965905 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-client-ca\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.966096 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-image-import-ca\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.966266 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.966444 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-service-ca\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.966530 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/15d7d52b-dc81-4400-bbce-353f00ff5103-config\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.966625 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1294227b-6b51-4269-93e2-542650c9dcb8-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.967112 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-etcd-serving-ca\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.967562 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/45e7c581-937a-46b2-9501-b704239b106c-audit-dir\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.968464 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7a1eb51-28c1-46db-ab15-7d01d75e42e9-config\") pod \"machine-approver-56656f9798-wmw8m\" (UID: \"c7a1eb51-28c1-46db-ab15-7d01d75e42e9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.968957 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15d7d52b-dc81-4400-bbce-353f00ff5103-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.969730 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1294227b-6b51-4269-93e2-542650c9dcb8-encryption-config\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.969749 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-oauth-config\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.969822 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.969843 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2k7k2"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.969852 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-w98w8"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.970179 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-config\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.970293 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.970346 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-config\") pod \"route-controller-manager-6576b87f9c-c5xml\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.970579 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f9ed5a87-d7c4-48bb-a331-afe0b8bc914a-config\") pod \"machine-api-operator-5694c8668f-6sk6v\" (UID: \"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.970830 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/15d7d52b-dc81-4400-bbce-353f00ff5103-service-ca-bundle\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.970855 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-audit-policies\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.971074 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/45e7c581-937a-46b2-9501-b704239b106c-encryption-config\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.971346 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-trusted-ca-bundle\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.971769 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-mpbsp"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.971790 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk"] Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.971821 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.971837 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1294227b-6b51-4269-93e2-542650c9dcb8-audit-policies\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.972053 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-trusted-ca-bundle\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.972414 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/c7a1eb51-28c1-46db-ab15-7d01d75e42e9-machine-approver-tls\") pod \"machine-approver-56656f9798-wmw8m\" (UID: \"c7a1eb51-28c1-46db-ab15-7d01d75e42e9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.972452 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/f9ed5a87-d7c4-48bb-a331-afe0b8bc914a-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-6sk6v\" (UID: \"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.972484 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.963789 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-audit-dir\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.972740 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-oauth-serving-cert\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.973172 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-audit\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.973961 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/f9ed5a87-d7c4-48bb-a331-afe0b8bc914a-images\") pod \"machine-api-operator-5694c8668f-6sk6v\" (UID: \"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.974034 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-client-ca\") pod \"route-controller-manager-6576b87f9c-c5xml\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.974076 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.974120 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/afb89818-ec4f-4a94-8a7a-39e0e798616a-available-featuregates\") pod \"openshift-config-operator-7777fb866f-d4ntr\" (UID: \"afb89818-ec4f-4a94-8a7a-39e0e798616a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.974653 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45e7c581-937a-46b2-9501-b704239b106c-config\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.974956 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/72724ed8-9cee-4f67-a055-c542e999b483-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-xl6pd\" (UID: \"72724ed8-9cee-4f67-a055-c542e999b483\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.975274 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-serving-cert\") pod \"route-controller-manager-6576b87f9c-c5xml\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.975343 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1294227b-6b51-4269-93e2-542650c9dcb8-audit-dir\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.975653 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.975732 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/45e7c581-937a-46b2-9501-b704239b106c-node-pullsecrets\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.975876 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.975906 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.975876 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/45e7c581-937a-46b2-9501-b704239b106c-etcd-client\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.976745 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.977244 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1294227b-6b51-4269-93e2-542650c9dcb8-etcd-client\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.977298 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afb89818-ec4f-4a94-8a7a-39e0e798616a-serving-cert\") pod \"openshift-config-operator-7777fb866f-d4ntr\" (UID: \"afb89818-ec4f-4a94-8a7a-39e0e798616a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.977392 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b86a263-c28c-4b03-af1e-1e57f4f203b7-serving-cert\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.977402 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/1c70671b-0773-4c66-bc16-cf7a3e383572-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-z48bm\" (UID: \"1c70671b-0773-4c66-bc16-cf7a3e383572\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.977478 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.978019 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.978107 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/15d7d52b-dc81-4400-bbce-353f00ff5103-serving-cert\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.978641 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1294227b-6b51-4269-93e2-542650c9dcb8-serving-cert\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.978679 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45e7c581-937a-46b2-9501-b704239b106c-serving-cert\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.978887 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-serving-cert\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.979088 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.979538 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:23 crc kubenswrapper[4930]: I0313 09:14:23.993574 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.007616 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.046587 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.066077 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/805de6fb-8159-4c0d-bba1-f2d20558fb40-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nk9n9\" (UID: \"805de6fb-8159-4c0d-bba1-f2d20558fb40\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.066138 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f920133-728b-43bb-a0ad-c851a2820b91-config\") pod \"kube-apiserver-operator-766d6c64bb-dq2fw\" (UID: \"5f920133-728b-43bb-a0ad-c851a2820b91\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.066177 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jg9lc\" (UniqueName: \"kubernetes.io/projected/805de6fb-8159-4c0d-bba1-f2d20558fb40-kube-api-access-jg9lc\") pod \"kube-storage-version-migrator-operator-b67b599dd-nk9n9\" (UID: \"805de6fb-8159-4c0d-bba1-f2d20558fb40\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.066217 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f920133-728b-43bb-a0ad-c851a2820b91-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-dq2fw\" (UID: \"5f920133-728b-43bb-a0ad-c851a2820b91\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.066292 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/805de6fb-8159-4c0d-bba1-f2d20558fb40-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nk9n9\" (UID: \"805de6fb-8159-4c0d-bba1-f2d20558fb40\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.066324 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33b20f7f-12af-43bd-8937-426f36818e3d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-r9svm\" (UID: \"33b20f7f-12af-43bd-8937-426f36818e3d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.066352 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ntjsh\" (UniqueName: \"kubernetes.io/projected/33b20f7f-12af-43bd-8937-426f36818e3d-kube-api-access-ntjsh\") pod \"openshift-controller-manager-operator-756b6f6bc6-r9svm\" (UID: \"33b20f7f-12af-43bd-8937-426f36818e3d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.066404 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33b20f7f-12af-43bd-8937-426f36818e3d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-r9svm\" (UID: \"33b20f7f-12af-43bd-8937-426f36818e3d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.066458 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f920133-728b-43bb-a0ad-c851a2820b91-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dq2fw\" (UID: \"5f920133-728b-43bb-a0ad-c851a2820b91\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.066773 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.066998 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/33b20f7f-12af-43bd-8937-426f36818e3d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-r9svm\" (UID: \"33b20f7f-12af-43bd-8937-426f36818e3d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.069964 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/33b20f7f-12af-43bd-8937-426f36818e3d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-r9svm\" (UID: \"33b20f7f-12af-43bd-8937-426f36818e3d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.076913 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5f920133-728b-43bb-a0ad-c851a2820b91-config\") pod \"kube-apiserver-operator-766d6c64bb-dq2fw\" (UID: \"5f920133-728b-43bb-a0ad-c851a2820b91\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.086649 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.099124 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5f920133-728b-43bb-a0ad-c851a2820b91-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dq2fw\" (UID: \"5f920133-728b-43bb-a0ad-c851a2820b91\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.106272 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.127060 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.146296 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.166933 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.186123 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.207198 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.226917 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.246884 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.266823 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.286672 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.307296 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.326173 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.346711 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.361916 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/805de6fb-8159-4c0d-bba1-f2d20558fb40-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-nk9n9\" (UID: \"805de6fb-8159-4c0d-bba1-f2d20558fb40\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.366801 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.377909 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/805de6fb-8159-4c0d-bba1-f2d20558fb40-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-nk9n9\" (UID: \"805de6fb-8159-4c0d-bba1-f2d20558fb40\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.386290 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.406629 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.426298 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.446891 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.467209 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.487046 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.514804 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.527659 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.547161 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.567443 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.622576 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.622890 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.626561 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.646938 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.666985 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.708808 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.727068 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.752093 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.766501 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.787391 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.807930 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.826966 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.847256 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.866924 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.884977 4930 request.go:700] Waited for 1.00275089s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-storage-version-migrator/configmaps?fieldSelector=metadata.name%3Dopenshift-service-ca.crt&limit=500&resourceVersion=0 Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.886641 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.907367 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.926860 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.946716 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.967425 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.969917 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.969929 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.970024 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.970193 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:24 crc kubenswrapper[4930]: I0313 09:14:24.986672 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.007270 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.026811 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.046493 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.067054 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.087147 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.107048 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.127777 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.146753 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.170208 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.186948 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.207408 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.227170 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.246507 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.266856 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.287572 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.306738 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.326721 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.346416 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.366426 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.386156 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.406830 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.426653 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.447207 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.467693 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.486873 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.507814 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.527233 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.546918 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.566649 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.586637 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.606173 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.626026 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.648337 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.666794 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.686300 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.706177 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.727594 4930 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.746311 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.766342 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.787691 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-sysctl-allowlist" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.839401 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h7w2\" (UniqueName: \"kubernetes.io/projected/1b86a263-c28c-4b03-af1e-1e57f4f203b7-kube-api-access-9h7w2\") pod \"controller-manager-879f6c89f-xqlzn\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.842153 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kttq\" (UniqueName: \"kubernetes.io/projected/15d7d52b-dc81-4400-bbce-353f00ff5103-kube-api-access-6kttq\") pod \"authentication-operator-69f744f599-9vpj8\" (UID: \"15d7d52b-dc81-4400-bbce-353f00ff5103\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.856785 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.865239 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jdpp\" (UniqueName: \"kubernetes.io/projected/72724ed8-9cee-4f67-a055-c542e999b483-kube-api-access-2jdpp\") pod \"cluster-samples-operator-665b6dd947-xl6pd\" (UID: \"72724ed8-9cee-4f67-a055-c542e999b483\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.880131 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.885837 4930 request.go:700] Waited for 1.914620471s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-machine-api/serviceaccounts/machine-api-operator/token Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.886064 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj7cw\" (UniqueName: \"kubernetes.io/projected/afb89818-ec4f-4a94-8a7a-39e0e798616a-kube-api-access-nj7cw\") pod \"openshift-config-operator-7777fb866f-d4ntr\" (UID: \"afb89818-ec4f-4a94-8a7a-39e0e798616a\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.907080 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qvsw\" (UniqueName: \"kubernetes.io/projected/f9ed5a87-d7c4-48bb-a331-afe0b8bc914a-kube-api-access-2qvsw\") pod \"machine-api-operator-5694c8668f-6sk6v\" (UID: \"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.913769 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.925176 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz52n\" (UniqueName: \"kubernetes.io/projected/45e7c581-937a-46b2-9501-b704239b106c-kube-api-access-lz52n\") pod \"apiserver-76f77b778f-f2kpv\" (UID: \"45e7c581-937a-46b2-9501-b704239b106c\") " pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.944281 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9srr9\" (UniqueName: \"kubernetes.io/projected/1c70671b-0773-4c66-bc16-cf7a3e383572-kube-api-access-9srr9\") pod \"cluster-image-registry-operator-dc59b4c8b-z48bm\" (UID: \"1c70671b-0773-4c66-bc16-cf7a3e383572\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:25 crc kubenswrapper[4930]: I0313 09:14:25.972522 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jwj6f\" (UniqueName: \"kubernetes.io/projected/c7a1eb51-28c1-46db-ab15-7d01d75e42e9-kube-api-access-jwj6f\") pod \"machine-approver-56656f9798-wmw8m\" (UID: \"c7a1eb51-28c1-46db-ab15-7d01d75e42e9\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.017754 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.019449 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.019814 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.023766 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngbzt\" (UniqueName: \"kubernetes.io/projected/1294227b-6b51-4269-93e2-542650c9dcb8-kube-api-access-ngbzt\") pod \"apiserver-7bbb656c7d-hjd4q\" (UID: \"1294227b-6b51-4269-93e2-542650c9dcb8\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.043591 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/1c70671b-0773-4c66-bc16-cf7a3e383572-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-z48bm\" (UID: \"1c70671b-0773-4c66-bc16-cf7a3e383572\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.045129 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.047877 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2668c\" (UniqueName: \"kubernetes.io/projected/0805c214-0e3a-44a0-8bfe-38e8adcccde4-kube-api-access-2668c\") pod \"console-f9d7485db-btdqm\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.053825 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htl4k\" (UniqueName: \"kubernetes.io/projected/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-kube-api-access-htl4k\") pod \"oauth-openshift-558db77b4-nsh7c\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:26 crc kubenswrapper[4930]: W0313 09:14:26.062881 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7a1eb51_28c1_46db_ab15_7d01d75e42e9.slice/crio-4201a493b4238d6cafa7fc6ce5759ec3ce40a7d9fe38315625660be419729d8a WatchSource:0}: Error finding container 4201a493b4238d6cafa7fc6ce5759ec3ce40a7d9fe38315625660be419729d8a: Status 404 returned error can't find the container with id 4201a493b4238d6cafa7fc6ce5759ec3ce40a7d9fe38315625660be419729d8a Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.064008 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjck8\" (UniqueName: \"kubernetes.io/projected/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-kube-api-access-fjck8\") pod \"route-controller-manager-6576b87f9c-c5xml\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.135290 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5f920133-728b-43bb-a0ad-c851a2820b91-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-dq2fw\" (UID: \"5f920133-728b-43bb-a0ad-c851a2820b91\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.142762 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.149142 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jg9lc\" (UniqueName: \"kubernetes.io/projected/805de6fb-8159-4c0d-bba1-f2d20558fb40-kube-api-access-jg9lc\") pod \"kube-storage-version-migrator-operator-b67b599dd-nk9n9\" (UID: \"805de6fb-8159-4c0d-bba1-f2d20558fb40\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.167044 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.173865 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ntjsh\" (UniqueName: \"kubernetes.io/projected/33b20f7f-12af-43bd-8937-426f36818e3d-kube-api-access-ntjsh\") pod \"openshift-controller-manager-operator-756b6f6bc6-r9svm\" (UID: \"33b20f7f-12af-43bd-8937-426f36818e3d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.190340 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.195743 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.207564 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.220950 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.227411 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.242739 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.248146 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.248931 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6tr92\" (UniqueName: \"kubernetes.io/projected/635a349e-a2f9-4fe3-b58a-72334ef95d6b-kube-api-access-6tr92\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249051 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf438604-87e1-45b2-8ba6-db4e0459dff9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-scftn\" (UID: \"bf438604-87e1-45b2-8ba6-db4e0459dff9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249127 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bfd7088d-61d7-4753-b25a-ea843f19eb30-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wrdm6\" (UID: \"bfd7088d-61d7-4753-b25a-ea843f19eb30\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249187 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-bound-sa-token\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249234 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7de6cb48-405d-43d2-8387-772f05f9d85b-etcd-service-ca\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249335 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249363 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wtfjf\" (UID: \"3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249398 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2feeb9b-9255-4660-829f-23c296a2a1c3-trusted-ca\") pod \"console-operator-58897d9998-w957v\" (UID: \"a2feeb9b-9255-4660-829f-23c296a2a1c3\") " pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249470 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf438604-87e1-45b2-8ba6-db4e0459dff9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-scftn\" (UID: \"bf438604-87e1-45b2-8ba6-db4e0459dff9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249504 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/635a349e-a2f9-4fe3-b58a-72334ef95d6b-service-ca-bundle\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249538 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/635a349e-a2f9-4fe3-b58a-72334ef95d6b-stats-auth\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249600 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/22f82073-0b04-4c7a-ab16-e88878214407-registry-certificates\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249623 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2feeb9b-9255-4660-829f-23c296a2a1c3-config\") pod \"console-operator-58897d9998-w957v\" (UID: \"a2feeb9b-9255-4660-829f-23c296a2a1c3\") " pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249656 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wtfjf\" (UID: \"3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249689 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/635a349e-a2f9-4fe3-b58a-72334ef95d6b-metrics-certs\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249711 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wt4bn\" (UniqueName: \"kubernetes.io/projected/dc201d97-a254-45f7-8f60-f2c5dd726852-kube-api-access-wt4bn\") pod \"downloads-7954f5f757-22ldf\" (UID: \"dc201d97-a254-45f7-8f60-f2c5dd726852\") " pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249772 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/95b6d230-5249-4b6f-a650-64de33200086-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rb28t\" (UID: \"95b6d230-5249-4b6f-a650-64de33200086\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249794 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfd7088d-61d7-4753-b25a-ea843f19eb30-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wrdm6\" (UID: \"bfd7088d-61d7-4753-b25a-ea843f19eb30\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249832 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/95b6d230-5249-4b6f-a650-64de33200086-metrics-tls\") pod \"ingress-operator-5b745b69d9-rb28t\" (UID: \"95b6d230-5249-4b6f-a650-64de33200086\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249852 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-registry-tls\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249896 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfd7088d-61d7-4753-b25a-ea843f19eb30-config\") pod \"kube-controller-manager-operator-78b949d7b-wrdm6\" (UID: \"bfd7088d-61d7-4753-b25a-ea843f19eb30\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249959 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf438604-87e1-45b2-8ba6-db4e0459dff9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-scftn\" (UID: \"bf438604-87e1-45b2-8ba6-db4e0459dff9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.249979 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2feeb9b-9255-4660-829f-23c296a2a1c3-serving-cert\") pod \"console-operator-58897d9998-w957v\" (UID: \"a2feeb9b-9255-4660-829f-23c296a2a1c3\") " pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.250003 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22f82073-0b04-4c7a-ab16-e88878214407-trusted-ca\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.250044 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qrxz\" (UniqueName: \"kubernetes.io/projected/7de6cb48-405d-43d2-8387-772f05f9d85b-kube-api-access-7qrxz\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.250104 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27gwq\" (UniqueName: \"kubernetes.io/projected/3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724-kube-api-access-27gwq\") pod \"openshift-apiserver-operator-796bbdcf4f-wtfjf\" (UID: \"3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.250140 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/635a349e-a2f9-4fe3-b58a-72334ef95d6b-default-certificate\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.250174 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7de6cb48-405d-43d2-8387-772f05f9d85b-serving-cert\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.250236 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7de6cb48-405d-43d2-8387-772f05f9d85b-etcd-ca\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.250271 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k472w\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-kube-api-access-k472w\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.250305 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95b6d230-5249-4b6f-a650-64de33200086-trusted-ca\") pod \"ingress-operator-5b745b69d9-rb28t\" (UID: \"95b6d230-5249-4b6f-a650-64de33200086\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.250350 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/22f82073-0b04-4c7a-ab16-e88878214407-installation-pull-secrets\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.250411 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7ssn\" (UniqueName: \"kubernetes.io/projected/95b6d230-5249-4b6f-a650-64de33200086-kube-api-access-w7ssn\") pod \"ingress-operator-5b745b69d9-rb28t\" (UID: \"95b6d230-5249-4b6f-a650-64de33200086\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.253339 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w8wm\" (UniqueName: \"kubernetes.io/projected/a2feeb9b-9255-4660-829f-23c296a2a1c3-kube-api-access-9w8wm\") pod \"console-operator-58897d9998-w957v\" (UID: \"a2feeb9b-9255-4660-829f-23c296a2a1c3\") " pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.253383 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/22f82073-0b04-4c7a-ab16-e88878214407-ca-trust-extracted\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.253463 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7de6cb48-405d-43d2-8387-772f05f9d85b-etcd-client\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.253528 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de6cb48-405d-43d2-8387-772f05f9d85b-config\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: E0313 09:14:26.254308 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:26.75429016 +0000 UTC m=+107.504204927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.267120 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.282267 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.304631 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.310041 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.341681 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.363770 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:26 crc kubenswrapper[4930]: E0313 09:14:26.363860 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:26.863840964 +0000 UTC m=+107.613755641 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.364278 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z72bw\" (UniqueName: \"kubernetes.io/projected/143f048a-f720-4d97-92ac-c9766a7e39ab-kube-api-access-z72bw\") pod \"catalog-operator-68c6474976-hrnlx\" (UID: \"143f048a-f720-4d97-92ac-c9766a7e39ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.364318 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8ea8f91d-30b3-452b-8aac-1085a5d58eab-node-bootstrap-token\") pod \"machine-config-server-rlvdx\" (UID: \"8ea8f91d-30b3-452b-8aac-1085a5d58eab\") " pod="openshift-machine-config-operator/machine-config-server-rlvdx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.364346 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf438604-87e1-45b2-8ba6-db4e0459dff9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-scftn\" (UID: \"bf438604-87e1-45b2-8ba6-db4e0459dff9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.364371 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5lfk\" (UniqueName: \"kubernetes.io/projected/8ea8f91d-30b3-452b-8aac-1085a5d58eab-kube-api-access-s5lfk\") pod \"machine-config-server-rlvdx\" (UID: \"8ea8f91d-30b3-452b-8aac-1085a5d58eab\") " pod="openshift-machine-config-operator/machine-config-server-rlvdx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.364393 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/95528b17-57d5-4044-8c19-48dd83168399-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-z6996\" (UID: \"95528b17-57d5-4044-8c19-48dd83168399\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.364417 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f65639e5-a600-44c3-a8b5-49aea65404e0-config-volume\") pod \"collect-profiles-29556540-w2lkz\" (UID: \"f65639e5-a600-44c3-a8b5-49aea65404e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365011 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bfd7088d-61d7-4753-b25a-ea843f19eb30-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wrdm6\" (UID: \"bfd7088d-61d7-4753-b25a-ea843f19eb30\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365048 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clwzg\" (UniqueName: \"kubernetes.io/projected/514985d1-67d6-4948-b7b5-69f7dabfb424-kube-api-access-clwzg\") pod \"dns-default-w98w8\" (UID: \"514985d1-67d6-4948-b7b5-69f7dabfb424\") " pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365090 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-bound-sa-token\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365113 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7de6cb48-405d-43d2-8387-772f05f9d85b-etcd-service-ca\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365135 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nw94\" (UniqueName: \"kubernetes.io/projected/5b3d4fa3-311d-4adf-9a68-57e6a5661009-kube-api-access-6nw94\") pod \"ingress-canary-mpbsp\" (UID: \"5b3d4fa3-311d-4adf-9a68-57e6a5661009\") " pod="openshift-ingress-canary/ingress-canary-mpbsp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365185 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/514985d1-67d6-4948-b7b5-69f7dabfb424-config-volume\") pod \"dns-default-w98w8\" (UID: \"514985d1-67d6-4948-b7b5-69f7dabfb424\") " pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365214 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f65639e5-a600-44c3-a8b5-49aea65404e0-secret-volume\") pod \"collect-profiles-29556540-w2lkz\" (UID: \"f65639e5-a600-44c3-a8b5-49aea65404e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365259 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365285 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4klzv\" (UniqueName: \"kubernetes.io/projected/d8df4458-d1a4-473a-8ec3-7f8051b4cd87-kube-api-access-4klzv\") pod \"migrator-59844c95c7-kvzkk\" (UID: \"d8df4458-d1a4-473a-8ec3-7f8051b4cd87\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kvzkk" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365312 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9mfz\" (UniqueName: \"kubernetes.io/projected/1b960305-f6c1-4892-bbc2-70cce9c21546-kube-api-access-r9mfz\") pod \"olm-operator-6b444d44fb-j8gcf\" (UID: \"1b960305-f6c1-4892-bbc2-70cce9c21546\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365340 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wtfjf\" (UID: \"3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365362 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2feeb9b-9255-4660-829f-23c296a2a1c3-trusted-ca\") pod \"console-operator-58897d9998-w957v\" (UID: \"a2feeb9b-9255-4660-829f-23c296a2a1c3\") " pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365396 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-ready\") pod \"cni-sysctl-allowlist-ds-qnw8j\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365417 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb104f34-0d73-40d9-9ba2-13401d61315d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rdbqj\" (UID: \"bb104f34-0d73-40d9-9ba2-13401d61315d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365488 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-mountpoint-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365518 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf438604-87e1-45b2-8ba6-db4e0459dff9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-scftn\" (UID: \"bf438604-87e1-45b2-8ba6-db4e0459dff9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365542 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4-images\") pod \"machine-config-operator-74547568cd-27rl4\" (UID: \"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365567 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/635a349e-a2f9-4fe3-b58a-72334ef95d6b-service-ca-bundle\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365588 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/514985d1-67d6-4948-b7b5-69f7dabfb424-metrics-tls\") pod \"dns-default-w98w8\" (UID: \"514985d1-67d6-4948-b7b5-69f7dabfb424\") " pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365610 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gr5rh\" (UniqueName: \"kubernetes.io/projected/c275e3c9-6bf5-4916-9ea5-e98baecacc87-kube-api-access-gr5rh\") pod \"service-ca-9c57cc56f-smklb\" (UID: \"c275e3c9-6bf5-4916-9ea5-e98baecacc87\") " pod="openshift-service-ca/service-ca-9c57cc56f-smklb" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365634 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/635a349e-a2f9-4fe3-b58a-72334ef95d6b-stats-auth\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365656 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/d60fb09e-fa42-482c-9228-d89352573e75-tmpfs\") pod \"packageserver-d55dfcdfc-8krq5\" (UID: \"d60fb09e-fa42-482c-9228-d89352573e75\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365676 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7q8l\" (UniqueName: \"kubernetes.io/projected/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-kube-api-access-f7q8l\") pod \"cni-sysctl-allowlist-ds-qnw8j\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365715 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/22f82073-0b04-4c7a-ab16-e88878214407-registry-certificates\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365737 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d60fb09e-fa42-482c-9228-d89352573e75-webhook-cert\") pod \"packageserver-d55dfcdfc-8krq5\" (UID: \"d60fb09e-fa42-482c-9228-d89352573e75\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365758 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hdcfs\" (UniqueName: \"kubernetes.io/projected/4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4-kube-api-access-hdcfs\") pod \"machine-config-operator-74547568cd-27rl4\" (UID: \"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365779 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/143f048a-f720-4d97-92ac-c9766a7e39ab-profile-collector-cert\") pod \"catalog-operator-68c6474976-hrnlx\" (UID: \"143f048a-f720-4d97-92ac-c9766a7e39ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365803 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2feeb9b-9255-4660-829f-23c296a2a1c3-config\") pod \"console-operator-58897d9998-w957v\" (UID: \"a2feeb9b-9255-4660-829f-23c296a2a1c3\") " pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365827 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tqnzq\" (UniqueName: \"kubernetes.io/projected/d60fb09e-fa42-482c-9228-d89352573e75-kube-api-access-tqnzq\") pod \"packageserver-d55dfcdfc-8krq5\" (UID: \"d60fb09e-fa42-482c-9228-d89352573e75\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365849 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e2e5b38e-1cee-49bf-a2c2-49919c061082-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-fl9px\" (UID: \"e2e5b38e-1cee-49bf-a2c2-49919c061082\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365871 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nwvg\" (UniqueName: \"kubernetes.io/projected/f65639e5-a600-44c3-a8b5-49aea65404e0-kube-api-access-4nwvg\") pod \"collect-profiles-29556540-w2lkz\" (UID: \"f65639e5-a600-44c3-a8b5-49aea65404e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365894 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b209dc6-6318-4983-ab60-569a294ad17f-serving-cert\") pod \"service-ca-operator-777779d784-496vm\" (UID: \"3b209dc6-6318-4983-ab60-569a294ad17f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365914 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxc5z\" (UniqueName: \"kubernetes.io/projected/e2e5b38e-1cee-49bf-a2c2-49919c061082-kube-api-access-dxc5z\") pod \"machine-config-controller-84d6567774-fl9px\" (UID: \"e2e5b38e-1cee-49bf-a2c2-49919c061082\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365937 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wtfjf\" (UID: \"3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365960 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4-proxy-tls\") pod \"machine-config-operator-74547568cd-27rl4\" (UID: \"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.365980 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1b960305-f6c1-4892-bbc2-70cce9c21546-srv-cert\") pod \"olm-operator-6b444d44fb-j8gcf\" (UID: \"1b960305-f6c1-4892-bbc2-70cce9c21546\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366002 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7sx5\" (UniqueName: \"kubernetes.io/projected/645348bb-55ca-4eb5-abb2-845f5a021f51-kube-api-access-j7sx5\") pod \"dns-operator-744455d44c-l2c8z\" (UID: \"645348bb-55ca-4eb5-abb2-845f5a021f51\") " pod="openshift-dns-operator/dns-operator-744455d44c-l2c8z" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366026 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8ea8f91d-30b3-452b-8aac-1085a5d58eab-certs\") pod \"machine-config-server-rlvdx\" (UID: \"8ea8f91d-30b3-452b-8aac-1085a5d58eab\") " pod="openshift-machine-config-operator/machine-config-server-rlvdx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366047 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-plugins-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366080 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/635a349e-a2f9-4fe3-b58a-72334ef95d6b-metrics-certs\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366102 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wt4bn\" (UniqueName: \"kubernetes.io/projected/dc201d97-a254-45f7-8f60-f2c5dd726852-kube-api-access-wt4bn\") pod \"downloads-7954f5f757-22ldf\" (UID: \"dc201d97-a254-45f7-8f60-f2c5dd726852\") " pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366127 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6px8q\" (UniqueName: \"kubernetes.io/projected/95528b17-57d5-4044-8c19-48dd83168399-kube-api-access-6px8q\") pod \"package-server-manager-789f6589d5-z6996\" (UID: \"95528b17-57d5-4044-8c19-48dd83168399\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366150 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stx2q\" (UniqueName: \"kubernetes.io/projected/bb104f34-0d73-40d9-9ba2-13401d61315d-kube-api-access-stx2q\") pod \"marketplace-operator-79b997595-rdbqj\" (UID: \"bb104f34-0d73-40d9-9ba2-13401d61315d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366187 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d60fb09e-fa42-482c-9228-d89352573e75-apiservice-cert\") pod \"packageserver-d55dfcdfc-8krq5\" (UID: \"d60fb09e-fa42-482c-9228-d89352573e75\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366207 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b209dc6-6318-4983-ab60-569a294ad17f-config\") pod \"service-ca-operator-777779d784-496vm\" (UID: \"3b209dc6-6318-4983-ab60-569a294ad17f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366229 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb104f34-0d73-40d9-9ba2-13401d61315d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rdbqj\" (UID: \"bb104f34-0d73-40d9-9ba2-13401d61315d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366251 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/95b6d230-5249-4b6f-a650-64de33200086-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rb28t\" (UID: \"95b6d230-5249-4b6f-a650-64de33200086\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366273 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-qnw8j\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366294 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/143f048a-f720-4d97-92ac-c9766a7e39ab-srv-cert\") pod \"catalog-operator-68c6474976-hrnlx\" (UID: \"143f048a-f720-4d97-92ac-c9766a7e39ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366318 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfd7088d-61d7-4753-b25a-ea843f19eb30-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wrdm6\" (UID: \"bfd7088d-61d7-4753-b25a-ea843f19eb30\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366340 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4-auth-proxy-config\") pod \"machine-config-operator-74547568cd-27rl4\" (UID: \"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366361 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nptzn\" (UID: \"d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nptzn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366409 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/95b6d230-5249-4b6f-a650-64de33200086-metrics-tls\") pod \"ingress-operator-5b745b69d9-rb28t\" (UID: \"95b6d230-5249-4b6f-a650-64de33200086\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366446 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-registry-tls\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366486 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfd7088d-61d7-4753-b25a-ea843f19eb30-config\") pod \"kube-controller-manager-operator-78b949d7b-wrdm6\" (UID: \"bfd7088d-61d7-4753-b25a-ea843f19eb30\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366531 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22f82073-0b04-4c7a-ab16-e88878214407-trusted-ca\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366553 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf438604-87e1-45b2-8ba6-db4e0459dff9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-scftn\" (UID: \"bf438604-87e1-45b2-8ba6-db4e0459dff9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366575 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2feeb9b-9255-4660-829f-23c296a2a1c3-serving-cert\") pod \"console-operator-58897d9998-w957v\" (UID: \"a2feeb9b-9255-4660-829f-23c296a2a1c3\") " pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366596 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-qnw8j\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366716 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/7de6cb48-405d-43d2-8387-772f05f9d85b-etcd-service-ca\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: E0313 09:14:26.368179 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:26.868164076 +0000 UTC m=+107.618078753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.366621 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c275e3c9-6bf5-4916-9ea5-e98baecacc87-signing-cabundle\") pod \"service-ca-9c57cc56f-smklb\" (UID: \"c275e3c9-6bf5-4916-9ea5-e98baecacc87\") " pod="openshift-service-ca/service-ca-9c57cc56f-smklb" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.368924 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-f2kpv"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.369645 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724-config\") pod \"openshift-apiserver-operator-796bbdcf4f-wtfjf\" (UID: \"3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.370966 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qrxz\" (UniqueName: \"kubernetes.io/projected/7de6cb48-405d-43d2-8387-772f05f9d85b-kube-api-access-7qrxz\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.371015 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c275e3c9-6bf5-4916-9ea5-e98baecacc87-signing-key\") pod \"service-ca-9c57cc56f-smklb\" (UID: \"c275e3c9-6bf5-4916-9ea5-e98baecacc87\") " pod="openshift-service-ca/service-ca-9c57cc56f-smklb" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.371070 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27gwq\" (UniqueName: \"kubernetes.io/projected/3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724-kube-api-access-27gwq\") pod \"openshift-apiserver-operator-796bbdcf4f-wtfjf\" (UID: \"3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.371103 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/635a349e-a2f9-4fe3-b58a-72334ef95d6b-default-certificate\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.371134 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6w9v\" (UniqueName: \"kubernetes.io/projected/3b209dc6-6318-4983-ab60-569a294ad17f-kube-api-access-j6w9v\") pod \"service-ca-operator-777779d784-496vm\" (UID: \"3b209dc6-6318-4983-ab60-569a294ad17f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.371159 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e2e5b38e-1cee-49bf-a2c2-49919c061082-proxy-tls\") pod \"machine-config-controller-84d6567774-fl9px\" (UID: \"e2e5b38e-1cee-49bf-a2c2-49919c061082\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.371186 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7e799ce-dd29-42a6-9895-8ec606537229-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rm7dk\" (UID: \"f7e799ce-dd29-42a6-9895-8ec606537229\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.371245 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7de6cb48-405d-43d2-8387-772f05f9d85b-serving-cert\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.371292 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7de6cb48-405d-43d2-8387-772f05f9d85b-etcd-ca\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.371314 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k472w\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-kube-api-access-k472w\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.371338 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-csi-data-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.383030 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/95b6d230-5249-4b6f-a650-64de33200086-metrics-tls\") pod \"ingress-operator-5b745b69d9-rb28t\" (UID: \"95b6d230-5249-4b6f-a650-64de33200086\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.383499 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/22f82073-0b04-4c7a-ab16-e88878214407-registry-certificates\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.383516 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-wtfjf\" (UID: \"3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.384192 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2feeb9b-9255-4660-829f-23c296a2a1c3-config\") pod \"console-operator-58897d9998-w957v\" (UID: \"a2feeb9b-9255-4660-829f-23c296a2a1c3\") " pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.384230 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a2feeb9b-9255-4660-829f-23c296a2a1c3-trusted-ca\") pod \"console-operator-58897d9998-w957v\" (UID: \"a2feeb9b-9255-4660-829f-23c296a2a1c3\") " pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.384263 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bf438604-87e1-45b2-8ba6-db4e0459dff9-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-scftn\" (UID: \"bf438604-87e1-45b2-8ba6-db4e0459dff9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.384721 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/635a349e-a2f9-4fe3-b58a-72334ef95d6b-service-ca-bundle\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.386133 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a2feeb9b-9255-4660-829f-23c296a2a1c3-serving-cert\") pod \"console-operator-58897d9998-w957v\" (UID: \"a2feeb9b-9255-4660-829f-23c296a2a1c3\") " pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.386632 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/635a349e-a2f9-4fe3-b58a-72334ef95d6b-metrics-certs\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.387398 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.387466 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-registry-tls\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.388311 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/7de6cb48-405d-43d2-8387-772f05f9d85b-etcd-ca\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.393175 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7de6cb48-405d-43d2-8387-772f05f9d85b-serving-cert\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.393221 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bfd7088d-61d7-4753-b25a-ea843f19eb30-config\") pod \"kube-controller-manager-operator-78b949d7b-wrdm6\" (UID: \"bfd7088d-61d7-4753-b25a-ea843f19eb30\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.395736 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bf438604-87e1-45b2-8ba6-db4e0459dff9-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-scftn\" (UID: \"bf438604-87e1-45b2-8ba6-db4e0459dff9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.397336 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22f82073-0b04-4c7a-ab16-e88878214407-trusted-ca\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.397603 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/635a349e-a2f9-4fe3-b58a-72334ef95d6b-stats-auth\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.398015 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/635a349e-a2f9-4fe3-b58a-72334ef95d6b-default-certificate\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.400913 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bfd7088d-61d7-4753-b25a-ea843f19eb30-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-wrdm6\" (UID: \"bfd7088d-61d7-4753-b25a-ea843f19eb30\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.403513 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95b6d230-5249-4b6f-a650-64de33200086-trusted-ca\") pod \"ingress-operator-5b745b69d9-rb28t\" (UID: \"95b6d230-5249-4b6f-a650-64de33200086\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.403574 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b3d4fa3-311d-4adf-9a68-57e6a5661009-cert\") pod \"ingress-canary-mpbsp\" (UID: \"5b3d4fa3-311d-4adf-9a68-57e6a5661009\") " pod="openshift-ingress-canary/ingress-canary-mpbsp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.403626 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/22f82073-0b04-4c7a-ab16-e88878214407-installation-pull-secrets\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.403681 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-registration-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.403732 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1b960305-f6c1-4892-bbc2-70cce9c21546-profile-collector-cert\") pod \"olm-operator-6b444d44fb-j8gcf\" (UID: \"1b960305-f6c1-4892-bbc2-70cce9c21546\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.403778 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gcj6\" (UniqueName: \"kubernetes.io/projected/f7e799ce-dd29-42a6-9895-8ec606537229-kube-api-access-9gcj6\") pod \"control-plane-machine-set-operator-78cbb6b69f-rm7dk\" (UID: \"f7e799ce-dd29-42a6-9895-8ec606537229\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.403882 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7ssn\" (UniqueName: \"kubernetes.io/projected/95b6d230-5249-4b6f-a650-64de33200086-kube-api-access-w7ssn\") pod \"ingress-operator-5b745b69d9-rb28t\" (UID: \"95b6d230-5249-4b6f-a650-64de33200086\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.403934 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slrgh\" (UniqueName: \"kubernetes.io/projected/4f1e0b58-9978-423f-84bc-f974dc6347b4-kube-api-access-slrgh\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.404078 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/22f82073-0b04-4c7a-ab16-e88878214407-ca-trust-extracted\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.404118 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7de6cb48-405d-43d2-8387-772f05f9d85b-etcd-client\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.404168 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w8wm\" (UniqueName: \"kubernetes.io/projected/a2feeb9b-9255-4660-829f-23c296a2a1c3-kube-api-access-9w8wm\") pod \"console-operator-58897d9998-w957v\" (UID: \"a2feeb9b-9255-4660-829f-23c296a2a1c3\") " pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.404197 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de6cb48-405d-43d2-8387-772f05f9d85b-config\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.404232 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6tr92\" (UniqueName: \"kubernetes.io/projected/635a349e-a2f9-4fe3-b58a-72334ef95d6b-kube-api-access-6tr92\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.404260 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/645348bb-55ca-4eb5-abb2-845f5a021f51-metrics-tls\") pod \"dns-operator-744455d44c-l2c8z\" (UID: \"645348bb-55ca-4eb5-abb2-845f5a021f51\") " pod="openshift-dns-operator/dns-operator-744455d44c-l2c8z" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.404288 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-socket-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.404317 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j7vj\" (UniqueName: \"kubernetes.io/projected/d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5-kube-api-access-2j7vj\") pod \"multus-admission-controller-857f4d67dd-nptzn\" (UID: \"d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nptzn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.408380 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/22f82073-0b04-4c7a-ab16-e88878214407-ca-trust-extracted\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.408661 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7de6cb48-405d-43d2-8387-772f05f9d85b-config\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.408914 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.408920 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-bound-sa-token\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.410241 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bfd7088d-61d7-4753-b25a-ea843f19eb30-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-wrdm6\" (UID: \"bfd7088d-61d7-4753-b25a-ea843f19eb30\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.410859 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.411239 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-9vpj8"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.418558 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/22f82073-0b04-4c7a-ab16-e88878214407-installation-pull-secrets\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.425996 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/7de6cb48-405d-43d2-8387-772f05f9d85b-etcd-client\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.426764 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xqlzn"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.426831 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-btdqm"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.428979 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/95b6d230-5249-4b6f-a650-64de33200086-trusted-ca\") pod \"ingress-operator-5b745b69d9-rb28t\" (UID: \"95b6d230-5249-4b6f-a650-64de33200086\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.435308 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.436200 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wt4bn\" (UniqueName: \"kubernetes.io/projected/dc201d97-a254-45f7-8f60-f2c5dd726852-kube-api-access-wt4bn\") pod \"downloads-7954f5f757-22ldf\" (UID: \"dc201d97-a254-45f7-8f60-f2c5dd726852\") " pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.453717 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/95b6d230-5249-4b6f-a650-64de33200086-bound-sa-token\") pod \"ingress-operator-5b745b69d9-rb28t\" (UID: \"95b6d230-5249-4b6f-a650-64de33200086\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.459472 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-6sk6v"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.470367 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd" event={"ID":"72724ed8-9cee-4f67-a055-c542e999b483","Type":"ContainerStarted","Data":"637218781d4dccd03c2550ee4c9f12ac95fbd2ed4420e97655be2a8c74d989be"} Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.472479 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" event={"ID":"1b86a263-c28c-4b03-af1e-1e57f4f203b7","Type":"ContainerStarted","Data":"dbe3e1ac734c173def970d40e3ebe38dff88c72c679d345f64762b5e8807344d"} Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.473755 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" event={"ID":"15d7d52b-dc81-4400-bbce-353f00ff5103","Type":"ContainerStarted","Data":"1e24d9accf53eb5b0ec814d812cdc1c8c9568d5d4c46613b79eb808fe397c920"} Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.474638 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" event={"ID":"c7a1eb51-28c1-46db-ab15-7d01d75e42e9","Type":"ContainerStarted","Data":"0527196d076e2da38329a43c672ceb1c5f1ccb45e7eff9eacded742d8cba3e9d"} Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.474684 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" event={"ID":"c7a1eb51-28c1-46db-ab15-7d01d75e42e9","Type":"ContainerStarted","Data":"4201a493b4238d6cafa7fc6ce5759ec3ce40a7d9fe38315625660be419729d8a"} Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.483177 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k472w\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-kube-api-access-k472w\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510113 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:26 crc kubenswrapper[4930]: E0313 09:14:26.510265 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.010230134 +0000 UTC m=+107.760144811 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510288 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-csi-data-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510333 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b3d4fa3-311d-4adf-9a68-57e6a5661009-cert\") pod \"ingress-canary-mpbsp\" (UID: \"5b3d4fa3-311d-4adf-9a68-57e6a5661009\") " pod="openshift-ingress-canary/ingress-canary-mpbsp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510352 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-registration-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510368 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gcj6\" (UniqueName: \"kubernetes.io/projected/f7e799ce-dd29-42a6-9895-8ec606537229-kube-api-access-9gcj6\") pod \"control-plane-machine-set-operator-78cbb6b69f-rm7dk\" (UID: \"f7e799ce-dd29-42a6-9895-8ec606537229\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510383 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1b960305-f6c1-4892-bbc2-70cce9c21546-profile-collector-cert\") pod \"olm-operator-6b444d44fb-j8gcf\" (UID: \"1b960305-f6c1-4892-bbc2-70cce9c21546\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510415 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slrgh\" (UniqueName: \"kubernetes.io/projected/4f1e0b58-9978-423f-84bc-f974dc6347b4-kube-api-access-slrgh\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510475 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/645348bb-55ca-4eb5-abb2-845f5a021f51-metrics-tls\") pod \"dns-operator-744455d44c-l2c8z\" (UID: \"645348bb-55ca-4eb5-abb2-845f5a021f51\") " pod="openshift-dns-operator/dns-operator-744455d44c-l2c8z" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510493 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-socket-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510529 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j7vj\" (UniqueName: \"kubernetes.io/projected/d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5-kube-api-access-2j7vj\") pod \"multus-admission-controller-857f4d67dd-nptzn\" (UID: \"d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nptzn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510549 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z72bw\" (UniqueName: \"kubernetes.io/projected/143f048a-f720-4d97-92ac-c9766a7e39ab-kube-api-access-z72bw\") pod \"catalog-operator-68c6474976-hrnlx\" (UID: \"143f048a-f720-4d97-92ac-c9766a7e39ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510566 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8ea8f91d-30b3-452b-8aac-1085a5d58eab-node-bootstrap-token\") pod \"machine-config-server-rlvdx\" (UID: \"8ea8f91d-30b3-452b-8aac-1085a5d58eab\") " pod="openshift-machine-config-operator/machine-config-server-rlvdx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510603 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5lfk\" (UniqueName: \"kubernetes.io/projected/8ea8f91d-30b3-452b-8aac-1085a5d58eab-kube-api-access-s5lfk\") pod \"machine-config-server-rlvdx\" (UID: \"8ea8f91d-30b3-452b-8aac-1085a5d58eab\") " pod="openshift-machine-config-operator/machine-config-server-rlvdx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510621 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/95528b17-57d5-4044-8c19-48dd83168399-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-z6996\" (UID: \"95528b17-57d5-4044-8c19-48dd83168399\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510637 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f65639e5-a600-44c3-a8b5-49aea65404e0-config-volume\") pod \"collect-profiles-29556540-w2lkz\" (UID: \"f65639e5-a600-44c3-a8b5-49aea65404e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510653 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clwzg\" (UniqueName: \"kubernetes.io/projected/514985d1-67d6-4948-b7b5-69f7dabfb424-kube-api-access-clwzg\") pod \"dns-default-w98w8\" (UID: \"514985d1-67d6-4948-b7b5-69f7dabfb424\") " pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510686 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nw94\" (UniqueName: \"kubernetes.io/projected/5b3d4fa3-311d-4adf-9a68-57e6a5661009-kube-api-access-6nw94\") pod \"ingress-canary-mpbsp\" (UID: \"5b3d4fa3-311d-4adf-9a68-57e6a5661009\") " pod="openshift-ingress-canary/ingress-canary-mpbsp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510702 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/514985d1-67d6-4948-b7b5-69f7dabfb424-config-volume\") pod \"dns-default-w98w8\" (UID: \"514985d1-67d6-4948-b7b5-69f7dabfb424\") " pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510718 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f65639e5-a600-44c3-a8b5-49aea65404e0-secret-volume\") pod \"collect-profiles-29556540-w2lkz\" (UID: \"f65639e5-a600-44c3-a8b5-49aea65404e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510736 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510782 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4klzv\" (UniqueName: \"kubernetes.io/projected/d8df4458-d1a4-473a-8ec3-7f8051b4cd87-kube-api-access-4klzv\") pod \"migrator-59844c95c7-kvzkk\" (UID: \"d8df4458-d1a4-473a-8ec3-7f8051b4cd87\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kvzkk" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510798 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9mfz\" (UniqueName: \"kubernetes.io/projected/1b960305-f6c1-4892-bbc2-70cce9c21546-kube-api-access-r9mfz\") pod \"olm-operator-6b444d44fb-j8gcf\" (UID: \"1b960305-f6c1-4892-bbc2-70cce9c21546\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510816 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-ready\") pod \"cni-sysctl-allowlist-ds-qnw8j\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510857 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb104f34-0d73-40d9-9ba2-13401d61315d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rdbqj\" (UID: \"bb104f34-0d73-40d9-9ba2-13401d61315d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510881 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-mountpoint-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510925 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4-images\") pod \"machine-config-operator-74547568cd-27rl4\" (UID: \"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510949 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/514985d1-67d6-4948-b7b5-69f7dabfb424-metrics-tls\") pod \"dns-default-w98w8\" (UID: \"514985d1-67d6-4948-b7b5-69f7dabfb424\") " pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.510969 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gr5rh\" (UniqueName: \"kubernetes.io/projected/c275e3c9-6bf5-4916-9ea5-e98baecacc87-kube-api-access-gr5rh\") pod \"service-ca-9c57cc56f-smklb\" (UID: \"c275e3c9-6bf5-4916-9ea5-e98baecacc87\") " pod="openshift-service-ca/service-ca-9c57cc56f-smklb" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511011 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/d60fb09e-fa42-482c-9228-d89352573e75-tmpfs\") pod \"packageserver-d55dfcdfc-8krq5\" (UID: \"d60fb09e-fa42-482c-9228-d89352573e75\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511033 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7q8l\" (UniqueName: \"kubernetes.io/projected/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-kube-api-access-f7q8l\") pod \"cni-sysctl-allowlist-ds-qnw8j\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511055 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d60fb09e-fa42-482c-9228-d89352573e75-webhook-cert\") pod \"packageserver-d55dfcdfc-8krq5\" (UID: \"d60fb09e-fa42-482c-9228-d89352573e75\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511095 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hdcfs\" (UniqueName: \"kubernetes.io/projected/4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4-kube-api-access-hdcfs\") pod \"machine-config-operator-74547568cd-27rl4\" (UID: \"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511116 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/143f048a-f720-4d97-92ac-c9766a7e39ab-profile-collector-cert\") pod \"catalog-operator-68c6474976-hrnlx\" (UID: \"143f048a-f720-4d97-92ac-c9766a7e39ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511138 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tqnzq\" (UniqueName: \"kubernetes.io/projected/d60fb09e-fa42-482c-9228-d89352573e75-kube-api-access-tqnzq\") pod \"packageserver-d55dfcdfc-8krq5\" (UID: \"d60fb09e-fa42-482c-9228-d89352573e75\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511183 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nwvg\" (UniqueName: \"kubernetes.io/projected/f65639e5-a600-44c3-a8b5-49aea65404e0-kube-api-access-4nwvg\") pod \"collect-profiles-29556540-w2lkz\" (UID: \"f65639e5-a600-44c3-a8b5-49aea65404e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511207 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e2e5b38e-1cee-49bf-a2c2-49919c061082-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-fl9px\" (UID: \"e2e5b38e-1cee-49bf-a2c2-49919c061082\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511250 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxc5z\" (UniqueName: \"kubernetes.io/projected/e2e5b38e-1cee-49bf-a2c2-49919c061082-kube-api-access-dxc5z\") pod \"machine-config-controller-84d6567774-fl9px\" (UID: \"e2e5b38e-1cee-49bf-a2c2-49919c061082\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511274 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b209dc6-6318-4983-ab60-569a294ad17f-serving-cert\") pod \"service-ca-operator-777779d784-496vm\" (UID: \"3b209dc6-6318-4983-ab60-569a294ad17f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511295 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4-proxy-tls\") pod \"machine-config-operator-74547568cd-27rl4\" (UID: \"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511336 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1b960305-f6c1-4892-bbc2-70cce9c21546-srv-cert\") pod \"olm-operator-6b444d44fb-j8gcf\" (UID: \"1b960305-f6c1-4892-bbc2-70cce9c21546\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511363 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8ea8f91d-30b3-452b-8aac-1085a5d58eab-certs\") pod \"machine-config-server-rlvdx\" (UID: \"8ea8f91d-30b3-452b-8aac-1085a5d58eab\") " pod="openshift-machine-config-operator/machine-config-server-rlvdx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511404 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-plugins-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511449 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7sx5\" (UniqueName: \"kubernetes.io/projected/645348bb-55ca-4eb5-abb2-845f5a021f51-kube-api-access-j7sx5\") pod \"dns-operator-744455d44c-l2c8z\" (UID: \"645348bb-55ca-4eb5-abb2-845f5a021f51\") " pod="openshift-dns-operator/dns-operator-744455d44c-l2c8z" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511475 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6px8q\" (UniqueName: \"kubernetes.io/projected/95528b17-57d5-4044-8c19-48dd83168399-kube-api-access-6px8q\") pod \"package-server-manager-789f6589d5-z6996\" (UID: \"95528b17-57d5-4044-8c19-48dd83168399\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511496 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stx2q\" (UniqueName: \"kubernetes.io/projected/bb104f34-0d73-40d9-9ba2-13401d61315d-kube-api-access-stx2q\") pod \"marketplace-operator-79b997595-rdbqj\" (UID: \"bb104f34-0d73-40d9-9ba2-13401d61315d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511491 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-csi-data-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511541 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d60fb09e-fa42-482c-9228-d89352573e75-apiservice-cert\") pod \"packageserver-d55dfcdfc-8krq5\" (UID: \"d60fb09e-fa42-482c-9228-d89352573e75\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511563 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b209dc6-6318-4983-ab60-569a294ad17f-config\") pod \"service-ca-operator-777779d784-496vm\" (UID: \"3b209dc6-6318-4983-ab60-569a294ad17f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511608 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb104f34-0d73-40d9-9ba2-13401d61315d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rdbqj\" (UID: \"bb104f34-0d73-40d9-9ba2-13401d61315d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511631 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/143f048a-f720-4d97-92ac-c9766a7e39ab-srv-cert\") pod \"catalog-operator-68c6474976-hrnlx\" (UID: \"143f048a-f720-4d97-92ac-c9766a7e39ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511654 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-qnw8j\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511698 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4-auth-proxy-config\") pod \"machine-config-operator-74547568cd-27rl4\" (UID: \"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511722 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nptzn\" (UID: \"d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nptzn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511782 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-qnw8j\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511805 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c275e3c9-6bf5-4916-9ea5-e98baecacc87-signing-cabundle\") pod \"service-ca-9c57cc56f-smklb\" (UID: \"c275e3c9-6bf5-4916-9ea5-e98baecacc87\") " pod="openshift-service-ca/service-ca-9c57cc56f-smklb" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511847 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c275e3c9-6bf5-4916-9ea5-e98baecacc87-signing-key\") pod \"service-ca-9c57cc56f-smklb\" (UID: \"c275e3c9-6bf5-4916-9ea5-e98baecacc87\") " pod="openshift-service-ca/service-ca-9c57cc56f-smklb" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511911 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6w9v\" (UniqueName: \"kubernetes.io/projected/3b209dc6-6318-4983-ab60-569a294ad17f-kube-api-access-j6w9v\") pod \"service-ca-operator-777779d784-496vm\" (UID: \"3b209dc6-6318-4983-ab60-569a294ad17f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511959 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e2e5b38e-1cee-49bf-a2c2-49919c061082-proxy-tls\") pod \"machine-config-controller-84d6567774-fl9px\" (UID: \"e2e5b38e-1cee-49bf-a2c2-49919c061082\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.511984 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7e799ce-dd29-42a6-9895-8ec606537229-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rm7dk\" (UID: \"f7e799ce-dd29-42a6-9895-8ec606537229\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.512114 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-registration-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.513637 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/514985d1-67d6-4948-b7b5-69f7dabfb424-config-volume\") pod \"dns-default-w98w8\" (UID: \"514985d1-67d6-4948-b7b5-69f7dabfb424\") " pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.515416 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/d60fb09e-fa42-482c-9228-d89352573e75-tmpfs\") pod \"packageserver-d55dfcdfc-8krq5\" (UID: \"d60fb09e-fa42-482c-9228-d89352573e75\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: E0313 09:14:26.515732 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.015716436 +0000 UTC m=+107.765631113 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.516525 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-ready\") pod \"cni-sysctl-allowlist-ds-qnw8j\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.522595 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d60fb09e-fa42-482c-9228-d89352573e75-webhook-cert\") pod \"packageserver-d55dfcdfc-8krq5\" (UID: \"d60fb09e-fa42-482c-9228-d89352573e75\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.522941 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1b960305-f6c1-4892-bbc2-70cce9c21546-profile-collector-cert\") pod \"olm-operator-6b444d44fb-j8gcf\" (UID: \"1b960305-f6c1-4892-bbc2-70cce9c21546\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.523995 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-socket-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.524677 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4-auth-proxy-config\") pod \"machine-config-operator-74547568cd-27rl4\" (UID: \"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.527144 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-mountpoint-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.527763 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4-images\") pod \"machine-config-operator-74547568cd-27rl4\" (UID: \"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.530139 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb104f34-0d73-40d9-9ba2-13401d61315d-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-rdbqj\" (UID: \"bb104f34-0d73-40d9-9ba2-13401d61315d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.531235 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.531294 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e2e5b38e-1cee-49bf-a2c2-49919c061082-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-fl9px\" (UID: \"e2e5b38e-1cee-49bf-a2c2-49919c061082\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.532734 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-cni-sysctl-allowlist\") pod \"cni-sysctl-allowlist-ds-qnw8j\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.533493 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/514985d1-67d6-4948-b7b5-69f7dabfb424-metrics-tls\") pod \"dns-default-w98w8\" (UID: \"514985d1-67d6-4948-b7b5-69f7dabfb424\") " pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.535729 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/5b3d4fa3-311d-4adf-9a68-57e6a5661009-cert\") pod \"ingress-canary-mpbsp\" (UID: \"5b3d4fa3-311d-4adf-9a68-57e6a5661009\") " pod="openshift-ingress-canary/ingress-canary-mpbsp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.535997 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1b960305-f6c1-4892-bbc2-70cce9c21546-srv-cert\") pod \"olm-operator-6b444d44fb-j8gcf\" (UID: \"1b960305-f6c1-4892-bbc2-70cce9c21546\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.536223 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27gwq\" (UniqueName: \"kubernetes.io/projected/3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724-kube-api-access-27gwq\") pod \"openshift-apiserver-operator-796bbdcf4f-wtfjf\" (UID: \"3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.536746 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-nptzn\" (UID: \"d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nptzn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.523998 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-tuning-conf-dir\") pod \"cni-sysctl-allowlist-ds-qnw8j\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.536883 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4f1e0b58-9978-423f-84bc-f974dc6347b4-plugins-dir\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.538060 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3b209dc6-6318-4983-ab60-569a294ad17f-serving-cert\") pod \"service-ca-operator-777779d784-496vm\" (UID: \"3b209dc6-6318-4983-ab60-569a294ad17f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.538604 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb104f34-0d73-40d9-9ba2-13401d61315d-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-rdbqj\" (UID: \"bb104f34-0d73-40d9-9ba2-13401d61315d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.539196 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/8ea8f91d-30b3-452b-8aac-1085a5d58eab-certs\") pod \"machine-config-server-rlvdx\" (UID: \"8ea8f91d-30b3-452b-8aac-1085a5d58eab\") " pod="openshift-machine-config-operator/machine-config-server-rlvdx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.539605 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f65639e5-a600-44c3-a8b5-49aea65404e0-secret-volume\") pod \"collect-profiles-29556540-w2lkz\" (UID: \"f65639e5-a600-44c3-a8b5-49aea65404e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.539853 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/645348bb-55ca-4eb5-abb2-845f5a021f51-metrics-tls\") pod \"dns-operator-744455d44c-l2c8z\" (UID: \"645348bb-55ca-4eb5-abb2-845f5a021f51\") " pod="openshift-dns-operator/dns-operator-744455d44c-l2c8z" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.539983 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e2e5b38e-1cee-49bf-a2c2-49919c061082-proxy-tls\") pod \"machine-config-controller-84d6567774-fl9px\" (UID: \"e2e5b38e-1cee-49bf-a2c2-49919c061082\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.540238 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qrxz\" (UniqueName: \"kubernetes.io/projected/7de6cb48-405d-43d2-8387-772f05f9d85b-kube-api-access-7qrxz\") pod \"etcd-operator-b45778765-xnlmh\" (UID: \"7de6cb48-405d-43d2-8387-772f05f9d85b\") " pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.543211 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3b209dc6-6318-4983-ab60-569a294ad17f-config\") pod \"service-ca-operator-777779d784-496vm\" (UID: \"3b209dc6-6318-4983-ab60-569a294ad17f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.543728 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/c275e3c9-6bf5-4916-9ea5-e98baecacc87-signing-cabundle\") pod \"service-ca-9c57cc56f-smklb\" (UID: \"c275e3c9-6bf5-4916-9ea5-e98baecacc87\") " pod="openshift-service-ca/service-ca-9c57cc56f-smklb" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.550953 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4-proxy-tls\") pod \"machine-config-operator-74547568cd-27rl4\" (UID: \"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.551666 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d60fb09e-fa42-482c-9228-d89352573e75-apiservice-cert\") pod \"packageserver-d55dfcdfc-8krq5\" (UID: \"d60fb09e-fa42-482c-9228-d89352573e75\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.554325 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/c275e3c9-6bf5-4916-9ea5-e98baecacc87-signing-key\") pod \"service-ca-9c57cc56f-smklb\" (UID: \"c275e3c9-6bf5-4916-9ea5-e98baecacc87\") " pod="openshift-service-ca/service-ca-9c57cc56f-smklb" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.557527 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/95528b17-57d5-4044-8c19-48dd83168399-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-z6996\" (UID: \"95528b17-57d5-4044-8c19-48dd83168399\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.558326 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/143f048a-f720-4d97-92ac-c9766a7e39ab-srv-cert\") pod \"catalog-operator-68c6474976-hrnlx\" (UID: \"143f048a-f720-4d97-92ac-c9766a7e39ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.559043 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/143f048a-f720-4d97-92ac-c9766a7e39ab-profile-collector-cert\") pod \"catalog-operator-68c6474976-hrnlx\" (UID: \"143f048a-f720-4d97-92ac-c9766a7e39ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.563554 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bf438604-87e1-45b2-8ba6-db4e0459dff9-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-scftn\" (UID: \"bf438604-87e1-45b2-8ba6-db4e0459dff9\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.570331 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w8wm\" (UniqueName: \"kubernetes.io/projected/a2feeb9b-9255-4660-829f-23c296a2a1c3-kube-api-access-9w8wm\") pod \"console-operator-58897d9998-w957v\" (UID: \"a2feeb9b-9255-4660-829f-23c296a2a1c3\") " pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.573377 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/8ea8f91d-30b3-452b-8aac-1085a5d58eab-node-bootstrap-token\") pod \"machine-config-server-rlvdx\" (UID: \"8ea8f91d-30b3-452b-8aac-1085a5d58eab\") " pod="openshift-machine-config-operator/machine-config-server-rlvdx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.575363 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/f7e799ce-dd29-42a6-9895-8ec606537229-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rm7dk\" (UID: \"f7e799ce-dd29-42a6-9895-8ec606537229\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.537653 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f65639e5-a600-44c3-a8b5-49aea65404e0-config-volume\") pod \"collect-profiles-29556540-w2lkz\" (UID: \"f65639e5-a600-44c3-a8b5-49aea65404e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.589282 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7ssn\" (UniqueName: \"kubernetes.io/projected/95b6d230-5249-4b6f-a650-64de33200086-kube-api-access-w7ssn\") pod \"ingress-operator-5b745b69d9-rb28t\" (UID: \"95b6d230-5249-4b6f-a650-64de33200086\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.601281 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6tr92\" (UniqueName: \"kubernetes.io/projected/635a349e-a2f9-4fe3-b58a-72334ef95d6b-kube-api-access-6tr92\") pod \"router-default-5444994796-gc8fp\" (UID: \"635a349e-a2f9-4fe3-b58a-72334ef95d6b\") " pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.617352 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:26 crc kubenswrapper[4930]: E0313 09:14:26.617776 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.117760904 +0000 UTC m=+107.867675581 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.631450 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nsh7c"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.646215 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nw94\" (UniqueName: \"kubernetes.io/projected/5b3d4fa3-311d-4adf-9a68-57e6a5661009-kube-api-access-6nw94\") pod \"ingress-canary-mpbsp\" (UID: \"5b3d4fa3-311d-4adf-9a68-57e6a5661009\") " pod="openshift-ingress-canary/ingress-canary-mpbsp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.648488 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.654856 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:26 crc kubenswrapper[4930]: W0313 09:14:26.658621 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d35e51a_2f8d_4da0_9b2e_0dcf2b29cd53.slice/crio-c311a63d469cc11d3965ebd3108f7335a282396d98fa9ec5436ffef560e9690f WatchSource:0}: Error finding container c311a63d469cc11d3965ebd3108f7335a282396d98fa9ec5436ffef560e9690f: Status 404 returned error can't find the container with id c311a63d469cc11d3965ebd3108f7335a282396d98fa9ec5436ffef560e9690f Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.660388 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.666096 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gcj6\" (UniqueName: \"kubernetes.io/projected/f7e799ce-dd29-42a6-9895-8ec606537229-kube-api-access-9gcj6\") pod \"control-plane-machine-set-operator-78cbb6b69f-rm7dk\" (UID: \"f7e799ce-dd29-42a6-9895-8ec606537229\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.685123 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slrgh\" (UniqueName: \"kubernetes.io/projected/4f1e0b58-9978-423f-84bc-f974dc6347b4-kube-api-access-slrgh\") pod \"csi-hostpathplugin-2k7k2\" (UID: \"4f1e0b58-9978-423f-84bc-f974dc6347b4\") " pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.702150 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4klzv\" (UniqueName: \"kubernetes.io/projected/d8df4458-d1a4-473a-8ec3-7f8051b4cd87-kube-api-access-4klzv\") pod \"migrator-59844c95c7-kvzkk\" (UID: \"d8df4458-d1a4-473a-8ec3-7f8051b4cd87\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kvzkk" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.717325 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.718671 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: E0313 09:14:26.719118 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.219096045 +0000 UTC m=+107.969010792 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.723334 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.731873 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.731958 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9mfz\" (UniqueName: \"kubernetes.io/projected/1b960305-f6c1-4892-bbc2-70cce9c21546-kube-api-access-r9mfz\") pod \"olm-operator-6b444d44fb-j8gcf\" (UID: \"1b960305-f6c1-4892-bbc2-70cce9c21546\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.742075 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.742388 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tqnzq\" (UniqueName: \"kubernetes.io/projected/d60fb09e-fa42-482c-9228-d89352573e75-kube-api-access-tqnzq\") pod \"packageserver-d55dfcdfc-8krq5\" (UID: \"d60fb09e-fa42-482c-9228-d89352573e75\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.748651 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.765917 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.779553 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kvzkk" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.780581 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7q8l\" (UniqueName: \"kubernetes.io/projected/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-kube-api-access-f7q8l\") pod \"cni-sysctl-allowlist-ds-qnw8j\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.786217 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.789988 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hdcfs\" (UniqueName: \"kubernetes.io/projected/4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4-kube-api-access-hdcfs\") pod \"machine-config-operator-74547568cd-27rl4\" (UID: \"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.792207 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.804382 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j7vj\" (UniqueName: \"kubernetes.io/projected/d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5-kube-api-access-2j7vj\") pod \"multus-admission-controller-857f4d67dd-nptzn\" (UID: \"d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-nptzn" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.810700 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.812274 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.819491 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:26 crc kubenswrapper[4930]: E0313 09:14:26.819662 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.319634824 +0000 UTC m=+108.069549501 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.819777 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:26 crc kubenswrapper[4930]: E0313 09:14:26.820075 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.320062646 +0000 UTC m=+108.069977323 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.822161 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z72bw\" (UniqueName: \"kubernetes.io/projected/143f048a-f720-4d97-92ac-c9766a7e39ab-kube-api-access-z72bw\") pod \"catalog-operator-68c6474976-hrnlx\" (UID: \"143f048a-f720-4d97-92ac-c9766a7e39ab\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.839888 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.845252 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.848896 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gr5rh\" (UniqueName: \"kubernetes.io/projected/c275e3c9-6bf5-4916-9ea5-e98baecacc87-kube-api-access-gr5rh\") pod \"service-ca-9c57cc56f-smklb\" (UID: \"c275e3c9-6bf5-4916-9ea5-e98baecacc87\") " pod="openshift-service-ca/service-ca-9c57cc56f-smklb" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.860753 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-mpbsp" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.862499 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.863280 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml"] Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.875743 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nwvg\" (UniqueName: \"kubernetes.io/projected/f65639e5-a600-44c3-a8b5-49aea65404e0-kube-api-access-4nwvg\") pod \"collect-profiles-29556540-w2lkz\" (UID: \"f65639e5-a600-44c3-a8b5-49aea65404e0\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.884861 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 09:14:26 crc kubenswrapper[4930]: W0313 09:14:26.898254 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1294227b_6b51_4269_93e2_542650c9dcb8.slice/crio-02e7c0a57450eb8813814aa97633fe4ec14298c2ce7465e9da310a1ebdc23377 WatchSource:0}: Error finding container 02e7c0a57450eb8813814aa97633fe4ec14298c2ce7465e9da310a1ebdc23377: Status 404 returned error can't find the container with id 02e7c0a57450eb8813814aa97633fe4ec14298c2ce7465e9da310a1ebdc23377 Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.898909 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxc5z\" (UniqueName: \"kubernetes.io/projected/e2e5b38e-1cee-49bf-a2c2-49919c061082-kube-api-access-dxc5z\") pod \"machine-config-controller-84d6567774-fl9px\" (UID: \"e2e5b38e-1cee-49bf-a2c2-49919c061082\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.901725 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6w9v\" (UniqueName: \"kubernetes.io/projected/3b209dc6-6318-4983-ab60-569a294ad17f-kube-api-access-j6w9v\") pod \"service-ca-operator-777779d784-496vm\" (UID: \"3b209dc6-6318-4983-ab60-569a294ad17f\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.896859 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.921617 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:26 crc kubenswrapper[4930]: E0313 09:14:26.922094 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.422052433 +0000 UTC m=+108.171967120 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.929610 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clwzg\" (UniqueName: \"kubernetes.io/projected/514985d1-67d6-4948-b7b5-69f7dabfb424-kube-api-access-clwzg\") pod \"dns-default-w98w8\" (UID: \"514985d1-67d6-4948-b7b5-69f7dabfb424\") " pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.946787 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7sx5\" (UniqueName: \"kubernetes.io/projected/645348bb-55ca-4eb5-abb2-845f5a021f51-kube-api-access-j7sx5\") pod \"dns-operator-744455d44c-l2c8z\" (UID: \"645348bb-55ca-4eb5-abb2-845f5a021f51\") " pod="openshift-dns-operator/dns-operator-744455d44c-l2c8z" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.981331 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5lfk\" (UniqueName: \"kubernetes.io/projected/8ea8f91d-30b3-452b-8aac-1085a5d58eab-kube-api-access-s5lfk\") pod \"machine-config-server-rlvdx\" (UID: \"8ea8f91d-30b3-452b-8aac-1085a5d58eab\") " pod="openshift-machine-config-operator/machine-config-server-rlvdx" Mar 13 09:14:26 crc kubenswrapper[4930]: I0313 09:14:26.983338 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6px8q\" (UniqueName: \"kubernetes.io/projected/95528b17-57d5-4044-8c19-48dd83168399-kube-api-access-6px8q\") pod \"package-server-manager-789f6589d5-z6996\" (UID: \"95528b17-57d5-4044-8c19-48dd83168399\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.006292 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stx2q\" (UniqueName: \"kubernetes.io/projected/bb104f34-0d73-40d9-9ba2-13401d61315d-kube-api-access-stx2q\") pod \"marketplace-operator-79b997595-rdbqj\" (UID: \"bb104f34-0d73-40d9-9ba2-13401d61315d\") " pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.024373 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:27 crc kubenswrapper[4930]: E0313 09:14:27.024746 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.524730278 +0000 UTC m=+108.274644955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.036876 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t"] Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.054107 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.059807 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-nptzn" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.066227 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.090126 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-kvzkk"] Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.099515 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-smklb" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.119725 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.124986 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:27 crc kubenswrapper[4930]: E0313 09:14:27.125482 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.625463273 +0000 UTC m=+108.375377950 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.125941 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-l2c8z" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.131693 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.145102 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.151706 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.167347 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rlvdx" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.173572 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.176954 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn"] Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.186949 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-xnlmh"] Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.227244 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:27 crc kubenswrapper[4930]: E0313 09:14:27.227584 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.727572404 +0000 UTC m=+108.477487081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.229423 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf"] Mar 13 09:14:27 crc kubenswrapper[4930]: W0313 09:14:27.261090 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7de6cb48_405d_43d2_8387_772f05f9d85b.slice/crio-e5077ddad7563cd12e26c9c0182661f6e54e9ad0d740bdbc54a656137664d07d WatchSource:0}: Error finding container e5077ddad7563cd12e26c9c0182661f6e54e9ad0d740bdbc54a656137664d07d: Status 404 returned error can't find the container with id e5077ddad7563cd12e26c9c0182661f6e54e9ad0d740bdbc54a656137664d07d Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.284633 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4"] Mar 13 09:14:27 crc kubenswrapper[4930]: W0313 09:14:27.298445 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1b960305_f6c1_4892_bbc2_70cce9c21546.slice/crio-82e415e9276990f8032f766d5297bdd957abbee04dffe5903aada3e504a8c5f0 WatchSource:0}: Error finding container 82e415e9276990f8032f766d5297bdd957abbee04dffe5903aada3e504a8c5f0: Status 404 returned error can't find the container with id 82e415e9276990f8032f766d5297bdd957abbee04dffe5903aada3e504a8c5f0 Mar 13 09:14:27 crc kubenswrapper[4930]: W0313 09:14:27.317283 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd8df4458_d1a4_473a_8ec3_7f8051b4cd87.slice/crio-5deebcb0eb23e2bfbedc37d9d95317ebf1ea5f8b011d6baf185eb5f24bfa3ba7 WatchSource:0}: Error finding container 5deebcb0eb23e2bfbedc37d9d95317ebf1ea5f8b011d6baf185eb5f24bfa3ba7: Status 404 returned error can't find the container with id 5deebcb0eb23e2bfbedc37d9d95317ebf1ea5f8b011d6baf185eb5f24bfa3ba7 Mar 13 09:14:27 crc kubenswrapper[4930]: W0313 09:14:27.317951 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5206b83e_eee7_4ad4_bba2_e0c306b84aaf.slice/crio-66ae88c2acc110e16791269c7861aa1ce230b4f99d990633a957594692689665 WatchSource:0}: Error finding container 66ae88c2acc110e16791269c7861aa1ce230b4f99d990633a957594692689665: Status 404 returned error can't find the container with id 66ae88c2acc110e16791269c7861aa1ce230b4f99d990633a957594692689665 Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.329592 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:27 crc kubenswrapper[4930]: E0313 09:14:27.329731 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.829708645 +0000 UTC m=+108.579623332 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.329921 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:27 crc kubenswrapper[4930]: E0313 09:14:27.330184 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.830176527 +0000 UTC m=+108.580091204 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.433701 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:27 crc kubenswrapper[4930]: E0313 09:14:27.434296 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:27.93428207 +0000 UTC m=+108.684196747 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.527154 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" event={"ID":"5206b83e-eee7-4ad4-bba2-e0c306b84aaf","Type":"ContainerStarted","Data":"66ae88c2acc110e16791269c7861aa1ce230b4f99d990633a957594692689665"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.538176 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-mpbsp"] Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.549687 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" event={"ID":"5f920133-728b-43bb-a0ad-c851a2820b91","Type":"ContainerStarted","Data":"e1d9f81cbd3858b63744bf2477b93d9d43a6dcedd228fa123e07cfc7c9f5163e"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.551680 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:27 crc kubenswrapper[4930]: E0313 09:14:27.552029 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.052013426 +0000 UTC m=+108.801928103 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.555214 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" event={"ID":"1c70671b-0773-4c66-bc16-cf7a3e383572","Type":"ContainerStarted","Data":"7c7ac0ad92381333495bbe1d0a7ed56cf877288ca3b8bb1617be7ca3a11b1346"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.555252 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" event={"ID":"1c70671b-0773-4c66-bc16-cf7a3e383572","Type":"ContainerStarted","Data":"8e59bffefbecd167dc3912f10f7977f8493b87f9d54da00e555964795ddfbc22"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.558404 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gc8fp" event={"ID":"635a349e-a2f9-4fe3-b58a-72334ef95d6b","Type":"ContainerStarted","Data":"dea8001e3cc4c766d3837faf01e5321b183e56d0ac974fe18b9b39d2acfa3186"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.560861 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" event={"ID":"95b6d230-5249-4b6f-a650-64de33200086","Type":"ContainerStarted","Data":"f95c67e20bd0d90434e91fca23c1b247b053d71e50b5e4c37ed06903cf0a987d"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.569337 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-2k7k2"] Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.571003 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-22ldf"] Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.571714 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" event={"ID":"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53","Type":"ContainerStarted","Data":"c311a63d469cc11d3965ebd3108f7335a282396d98fa9ec5436ffef560e9690f"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.576048 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" event={"ID":"1b960305-f6c1-4892-bbc2-70cce9c21546","Type":"ContainerStarted","Data":"82e415e9276990f8032f766d5297bdd957abbee04dffe5903aada3e504a8c5f0"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.582855 4930 generic.go:334] "Generic (PLEG): container finished" podID="45e7c581-937a-46b2-9501-b704239b106c" containerID="44a8a2fbc8faeb932c9ef90455ac04b8cba9b38d15bdd802a40f17825902d074" exitCode=0 Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.582944 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" event={"ID":"45e7c581-937a-46b2-9501-b704239b106c","Type":"ContainerDied","Data":"44a8a2fbc8faeb932c9ef90455ac04b8cba9b38d15bdd802a40f17825902d074"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.582966 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" event={"ID":"45e7c581-937a-46b2-9501-b704239b106c","Type":"ContainerStarted","Data":"5c637e555096c969557fd3ed06e6c4b8250be54cdcf21c9f1da53647ae0099fd"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.584628 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" event={"ID":"1b86a263-c28c-4b03-af1e-1e57f4f203b7","Type":"ContainerStarted","Data":"b6f73c03b04c6dfbedf2131baf5f9683ec7d3c8144448da982246657b98af284"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.585336 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.587759 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" event={"ID":"bf438604-87e1-45b2-8ba6-db4e0459dff9","Type":"ContainerStarted","Data":"536d5f8139d2038879d9574150c7dee47fdf2c1686509954e11bb7856b176ba2"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.590085 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" event={"ID":"bfd7088d-61d7-4753-b25a-ea843f19eb30","Type":"ContainerStarted","Data":"a8f4d0cdddc6894b0a2b76db7b72aeb2611618df3754ee15917d62c890aac2d2"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.592505 4930 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-xqlzn container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.592539 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" podUID="1b86a263-c28c-4b03-af1e-1e57f4f203b7" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.592907 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" event={"ID":"7de6cb48-405d-43d2-8387-772f05f9d85b","Type":"ContainerStarted","Data":"e5077ddad7563cd12e26c9c0182661f6e54e9ad0d740bdbc54a656137664d07d"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.594070 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" event={"ID":"33b20f7f-12af-43bd-8937-426f36818e3d","Type":"ContainerStarted","Data":"c9758c1c4230cc63f31c3dd3eca6b9436b0d20f7ebd1c1de163e6db4eafb3fb7"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.595916 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" event={"ID":"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a","Type":"ContainerStarted","Data":"b7b7031dffbdebe0786563e16ab3293715fbae37234e0ad32a7a6c65c47ab90e"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.595940 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" event={"ID":"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a","Type":"ContainerStarted","Data":"a000dea3a22a400d340221b960e29151e742260bef128ab436c55ba66ad7f57e"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.597004 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" event={"ID":"c7a1eb51-28c1-46db-ab15-7d01d75e42e9","Type":"ContainerStarted","Data":"fdfc108bbd3281054f211fb4fc277ebe966b31b719860f43f36e315aba044ef2"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.597755 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kvzkk" event={"ID":"d8df4458-d1a4-473a-8ec3-7f8051b4cd87","Type":"ContainerStarted","Data":"5deebcb0eb23e2bfbedc37d9d95317ebf1ea5f8b011d6baf185eb5f24bfa3ba7"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.598612 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" event={"ID":"04a63d8d-f300-4a8b-9ba3-5093bb7292cb","Type":"ContainerStarted","Data":"ccde838ac1d4a7709450a0bd8d9fd6947c3a0001a5607d6f258caba63ce89d40"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.600811 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-btdqm" event={"ID":"0805c214-0e3a-44a0-8bfe-38e8adcccde4","Type":"ContainerStarted","Data":"09b323c026cfd6d4a081277c5fdd13281297e57f7b3186ed70554ccd759c673a"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.600836 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-btdqm" event={"ID":"0805c214-0e3a-44a0-8bfe-38e8adcccde4","Type":"ContainerStarted","Data":"5671701f6deb4c53d9a9f0b7c2b92122f6822cbe62c08a424d4da663d1002859"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.602761 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" event={"ID":"15d7d52b-dc81-4400-bbce-353f00ff5103","Type":"ContainerStarted","Data":"f8eac0b1a2476de135d617d21cdcefaae2507867e2a9d881bb3deb2366d37e5b"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.605363 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd" event={"ID":"72724ed8-9cee-4f67-a055-c542e999b483","Type":"ContainerStarted","Data":"85024d053f7e8ff562cec07cd305ac5b070e4be95a20502948475d034f6b04ba"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.605418 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd" event={"ID":"72724ed8-9cee-4f67-a055-c542e999b483","Type":"ContainerStarted","Data":"02cd074298d7503b8f17180f16a911b86452615cb3077d8de5b5f81ed1fe5f7f"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.606514 4930 generic.go:334] "Generic (PLEG): container finished" podID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerID="0a5cc9e28ce14c23e3294f7be876edef87f7364c3436206b4093befe012bf81a" exitCode=0 Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.606629 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" event={"ID":"afb89818-ec4f-4a94-8a7a-39e0e798616a","Type":"ContainerDied","Data":"0a5cc9e28ce14c23e3294f7be876edef87f7364c3436206b4093befe012bf81a"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.606662 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" event={"ID":"afb89818-ec4f-4a94-8a7a-39e0e798616a","Type":"ContainerStarted","Data":"bb9d08197e78ea9e5194cd70a27bcc70da2fcd21b263e0e8762dfe0383ac51c3"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.607550 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" event={"ID":"805de6fb-8159-4c0d-bba1-f2d20558fb40","Type":"ContainerStarted","Data":"8be976d2380be9f722fad59b1e9e57694ee4f416be8907e96f066b61680f29c5"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.609362 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" event={"ID":"1294227b-6b51-4269-93e2-542650c9dcb8","Type":"ContainerStarted","Data":"02e7c0a57450eb8813814aa97633fe4ec14298c2ce7465e9da310a1ebdc23377"} Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.653836 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:27 crc kubenswrapper[4930]: E0313 09:14:27.655851 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.155827881 +0000 UTC m=+108.905742568 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.716702 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=1.716685981 podStartE2EDuration="1.716685981s" podCreationTimestamp="2026-03-13 09:14:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:27.715170701 +0000 UTC m=+108.465085378" watchObservedRunningTime="2026-03-13 09:14:27.716685981 +0000 UTC m=+108.466600658" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.755539 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:27 crc kubenswrapper[4930]: E0313 09:14:27.756762 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.256652908 +0000 UTC m=+109.006567675 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.857455 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:27 crc kubenswrapper[4930]: E0313 09:14:27.863422 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.363393839 +0000 UTC m=+109.113308516 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:27 crc kubenswrapper[4930]: E0313 09:14:27.882802 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.382786882 +0000 UTC m=+109.132701559 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.895077 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:27 crc kubenswrapper[4930]: I0313 09:14:27.995947 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:27 crc kubenswrapper[4930]: E0313 09:14:27.998353 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.498329092 +0000 UTC m=+109.248243769 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.099927 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.100028 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs\") pod \"network-metrics-daemon-vfjf8\" (UID: \"8298eba9-72fc-4be8-9736-defaad400dcd\") " pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:28 crc kubenswrapper[4930]: E0313 09:14:28.100522 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.600502304 +0000 UTC m=+109.350417081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.111851 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8298eba9-72fc-4be8-9736-defaad400dcd-metrics-certs\") pod \"network-metrics-daemon-vfjf8\" (UID: \"8298eba9-72fc-4be8-9736-defaad400dcd\") " pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.162172 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-w957v"] Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.162539 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf"] Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.200999 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:28 crc kubenswrapper[4930]: E0313 09:14:28.201179 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.701151447 +0000 UTC m=+109.451066124 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.201265 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:28 crc kubenswrapper[4930]: E0313 09:14:28.201576 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.701563628 +0000 UTC m=+109.451478295 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.236108 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk"] Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.258297 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx"] Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.271480 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-496vm"] Mar 13 09:14:28 crc kubenswrapper[4930]: W0313 09:14:28.274291 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b209dc6_6318_4983_ab60_569a294ad17f.slice/crio-09f75a72a3af9852d46fb94748350200492b3d4ccbe9067dff5f5e12ce034abc WatchSource:0}: Error finding container 09f75a72a3af9852d46fb94748350200492b3d4ccbe9067dff5f5e12ce034abc: Status 404 returned error can't find the container with id 09f75a72a3af9852d46fb94748350200492b3d4ccbe9067dff5f5e12ce034abc Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.276292 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5"] Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.302006 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:28 crc kubenswrapper[4930]: E0313 09:14:28.303978 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.803952745 +0000 UTC m=+109.553867412 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.306129 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:28 crc kubenswrapper[4930]: E0313 09:14:28.307307 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.806656236 +0000 UTC m=+109.556570913 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.327169 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-vfjf8" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.406731 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:28 crc kubenswrapper[4930]: E0313 09:14:28.407109 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:28.907094713 +0000 UTC m=+109.657009390 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.516304 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:28 crc kubenswrapper[4930]: E0313 09:14:28.516699 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:29.016669717 +0000 UTC m=+109.766584394 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.626477 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:28 crc kubenswrapper[4930]: E0313 09:14:28.626803 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:29.126787476 +0000 UTC m=+109.876702153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.630179 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz"] Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.641009 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-w98w8"] Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.643453 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdbqj"] Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.663200 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rlvdx" event={"ID":"8ea8f91d-30b3-452b-8aac-1085a5d58eab","Type":"ContainerStarted","Data":"83a60e6065873cc8902fb5c7d29c52061929229d0d9b6449d1ec163579ebf25f"} Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.664052 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" podStartSLOduration=62.664034763 podStartE2EDuration="1m2.664034763s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:28.661884217 +0000 UTC m=+109.411798894" watchObservedRunningTime="2026-03-13 09:14:28.664034763 +0000 UTC m=+109.413949440" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.671823 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" event={"ID":"1b960305-f6c1-4892-bbc2-70cce9c21546","Type":"ContainerStarted","Data":"7884531756cb3156917f38dc87d7796a7de383a57b50e29e1f711951f47dea05"} Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.672718 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.676672 4930 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-j8gcf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.676707 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" podUID="1b960305-f6c1-4892-bbc2-70cce9c21546" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.684099 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-smklb"] Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.696993 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" event={"ID":"bf438604-87e1-45b2-8ba6-db4e0459dff9","Type":"ContainerStarted","Data":"2eab6a0c92b87a17b8a55e7867c70855257d9b0e28d17c88950a5ce73d6fb17a"} Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.716828 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996"] Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.728042 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:28 crc kubenswrapper[4930]: E0313 09:14:28.728391 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:29.228378623 +0000 UTC m=+109.978293290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.747231 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" podStartSLOduration=62.747212902 podStartE2EDuration="1m2.747212902s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:28.725026326 +0000 UTC m=+109.474941003" watchObservedRunningTime="2026-03-13 09:14:28.747212902 +0000 UTC m=+109.497127579" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.772664 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" event={"ID":"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53","Type":"ContainerStarted","Data":"d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124"} Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.773137 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.807885 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-btdqm" podStartSLOduration=62.807869866 podStartE2EDuration="1m2.807869866s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:28.74637304 +0000 UTC m=+109.496287717" watchObservedRunningTime="2026-03-13 09:14:28.807869866 +0000 UTC m=+109.557784543" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.809099 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.835702 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:28 crc kubenswrapper[4930]: E0313 09:14:28.835991 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:29.335944325 +0000 UTC m=+110.085859002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.836069 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:28 crc kubenswrapper[4930]: E0313 09:14:28.837979 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:29.337970728 +0000 UTC m=+110.087885405 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.868684 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" event={"ID":"33b20f7f-12af-43bd-8937-426f36818e3d","Type":"ContainerStarted","Data":"e0d725bcd7867f0df3a49877ffcdcdbad5026566a7eade8d2feb10aef9e6d52e"} Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.870646 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-xl6pd" podStartSLOduration=62.870638506 podStartE2EDuration="1m2.870638506s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:28.869833175 +0000 UTC m=+109.619747852" watchObservedRunningTime="2026-03-13 09:14:28.870638506 +0000 UTC m=+109.620553183" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.870814 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-z48bm" podStartSLOduration=62.87080941 podStartE2EDuration="1m2.87080941s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:28.808099662 +0000 UTC m=+109.558014339" watchObservedRunningTime="2026-03-13 09:14:28.87080941 +0000 UTC m=+109.620724087" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.894805 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-vfjf8"] Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.903679 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" event={"ID":"3b209dc6-6318-4983-ab60-569a294ad17f","Type":"ContainerStarted","Data":"09f75a72a3af9852d46fb94748350200492b3d4ccbe9067dff5f5e12ce034abc"} Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.913153 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px"] Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.926844 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" event={"ID":"5f920133-728b-43bb-a0ad-c851a2820b91","Type":"ContainerStarted","Data":"06ae8f962f661f8b2535708e3fa8e0555762c293343ee92cf59be2af1118fb05"} Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.938066 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:28 crc kubenswrapper[4930]: E0313 09:14:28.939732 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:29.439715349 +0000 UTC m=+110.189630026 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.944615 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" event={"ID":"d60fb09e-fa42-482c-9228-d89352573e75","Type":"ContainerStarted","Data":"6bfa27481fdd6a14a54c93aca4b6a13d5febafe5df753f97410659204aa4183f"} Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.955587 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-wmw8m" podStartSLOduration=62.955571061 podStartE2EDuration="1m2.955571061s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:28.953973989 +0000 UTC m=+109.703888666" watchObservedRunningTime="2026-03-13 09:14:28.955571061 +0000 UTC m=+109.705485738" Mar 13 09:14:28 crc kubenswrapper[4930]: I0313 09:14:28.958662 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-l2c8z"] Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.026574 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" event={"ID":"805de6fb-8159-4c0d-bba1-f2d20558fb40","Type":"ContainerStarted","Data":"030af6db881d0f9eb17ef2b9dffa1fabeefb261936b76c74874156aa84de514f"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.039227 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:29 crc kubenswrapper[4930]: E0313 09:14:29.039491 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:29.539480259 +0000 UTC m=+110.289394926 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.047757 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-nptzn"] Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.055399 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-22ldf" event={"ID":"dc201d97-a254-45f7-8f60-f2c5dd726852","Type":"ContainerStarted","Data":"70b080d5cce4f1eb0f2fb93722ef0e423b3caa36bd8f0f802317de127f7a588e"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.056737 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.064058 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.064113 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.075355 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" event={"ID":"04a63d8d-f300-4a8b-9ba3-5093bb7292cb","Type":"ContainerStarted","Data":"cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.076300 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.099565 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk" event={"ID":"f7e799ce-dd29-42a6-9895-8ec606537229","Type":"ContainerStarted","Data":"42105075c74c314e29ef1d48eb07a02fc6382b93f8782df1ebf3a9bba81dfc42"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.108317 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" podStartSLOduration=63.108303135 podStartE2EDuration="1m3.108303135s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.10732678 +0000 UTC m=+109.857241457" watchObservedRunningTime="2026-03-13 09:14:29.108303135 +0000 UTC m=+109.858217812" Mar 13 09:14:29 crc kubenswrapper[4930]: W0313 09:14:29.127003 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd3f426eb_df52_49eb_b7e2_be9ac0b9bfb5.slice/crio-a2c274a8a7bc2d874e84e16482149e1d065b5730391c352b3ac0a4a722fa4340 WatchSource:0}: Error finding container a2c274a8a7bc2d874e84e16482149e1d065b5730391c352b3ac0a4a722fa4340: Status 404 returned error can't find the container with id a2c274a8a7bc2d874e84e16482149e1d065b5730391c352b3ac0a4a722fa4340 Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.129920 4930 ???:1] "http: TLS handshake error from 192.168.126.11:56182: no serving certificate available for the kubelet" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.130264 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" event={"ID":"3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724","Type":"ContainerStarted","Data":"8f06a754dd32241d7e4ca8182f093c6533ca5ebabe1527b769017a324ed92355"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.139824 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:29 crc kubenswrapper[4930]: E0313 09:14:29.140778 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:29.640764058 +0000 UTC m=+110.390678735 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.141405 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-mpbsp" event={"ID":"5b3d4fa3-311d-4adf-9a68-57e6a5661009","Type":"ContainerStarted","Data":"786dc03b99fa030661bc89452d047ed369971bc634e92a261f2d93c470a60c64"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.154582 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-scftn" podStartSLOduration=62.154562666 podStartE2EDuration="1m2.154562666s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.147304528 +0000 UTC m=+109.897219205" watchObservedRunningTime="2026-03-13 09:14:29.154562666 +0000 UTC m=+109.904477343" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.188683 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" event={"ID":"f9ed5a87-d7c4-48bb-a331-afe0b8bc914a","Type":"ContainerStarted","Data":"1527a5b876fcccaea40f227fb511307cd95598328327328a51b592c3849dbf21"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.193329 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-r9svm" podStartSLOduration=63.193314372 podStartE2EDuration="1m3.193314372s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.193057765 +0000 UTC m=+109.942972442" watchObservedRunningTime="2026-03-13 09:14:29.193314372 +0000 UTC m=+109.943229049" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.227896 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" event={"ID":"afb89818-ec4f-4a94-8a7a-39e0e798616a","Type":"ContainerStarted","Data":"96e5faa3c3871a454979909e86e16277494094e8bfbd8123b70f92326e00fa79"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.228398 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.243809 4930 ???:1] "http: TLS handshake error from 192.168.126.11:56194: no serving certificate available for the kubelet" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.244697 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.245392 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dq2fw" podStartSLOduration=62.245375264 podStartE2EDuration="1m2.245375264s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.245196199 +0000 UTC m=+109.995110876" watchObservedRunningTime="2026-03-13 09:14:29.245375264 +0000 UTC m=+109.995289941" Mar 13 09:14:29 crc kubenswrapper[4930]: E0313 09:14:29.247255 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:29.747240192 +0000 UTC m=+110.497154869 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.272986 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" event={"ID":"143f048a-f720-4d97-92ac-c9766a7e39ab","Type":"ContainerStarted","Data":"86b6bbbd2ddb909ef54004412c0dbea16baade737972577f79592d4d4d9912e8"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.273864 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.281756 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-w957v" event={"ID":"a2feeb9b-9255-4660-829f-23c296a2a1c3","Type":"ContainerStarted","Data":"73fb43b4e3932026a362947fbf9a57830d680b97b953b5403dc3a5c6cbff9bf3"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.282517 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.285599 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.285646 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.298340 4930 generic.go:334] "Generic (PLEG): container finished" podID="1294227b-6b51-4269-93e2-542650c9dcb8" containerID="a13f2a95ad602dc99913d9557aefe4ccce145db34c0a46a21320b53a5a5420d3" exitCode=0 Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.298452 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" event={"ID":"1294227b-6b51-4269-93e2-542650c9dcb8","Type":"ContainerDied","Data":"a13f2a95ad602dc99913d9557aefe4ccce145db34c0a46a21320b53a5a5420d3"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.305710 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.305749 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.305858 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" event={"ID":"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4","Type":"ContainerStarted","Data":"1e48af6e521eae16f691d7f42953ad90d0ccd5d6f317f535105a3c6eca1a8696"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.305882 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" event={"ID":"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4","Type":"ContainerStarted","Data":"34c0a8472cb4f2ce3b2409dc6d527b477c39d9f98d765249f8fc0b5049d714ef"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.320355 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kvzkk" event={"ID":"d8df4458-d1a4-473a-8ec3-7f8051b4cd87","Type":"ContainerStarted","Data":"6b0e7ba751451dca873c4bba1d274b29f2752b3e2d536ff17fc14e694cb26fad"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.321842 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" podStartSLOduration=6.321832318 podStartE2EDuration="6.321832318s" podCreationTimestamp="2026-03-13 09:14:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.292874477 +0000 UTC m=+110.042789164" watchObservedRunningTime="2026-03-13 09:14:29.321832318 +0000 UTC m=+110.071746995" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.323835 4930 ???:1] "http: TLS handshake error from 192.168.126.11:56196: no serving certificate available for the kubelet" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.329092 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" event={"ID":"bfd7088d-61d7-4753-b25a-ea843f19eb30","Type":"ContainerStarted","Data":"ec076671f43e639665487f0ffaf9c2c7e3310805663513cc848268fdcf45f681"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.333380 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" podStartSLOduration=62.333347197 podStartE2EDuration="1m2.333347197s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.324091227 +0000 UTC m=+110.074005904" watchObservedRunningTime="2026-03-13 09:14:29.333347197 +0000 UTC m=+110.083261874" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.333495 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" event={"ID":"4f1e0b58-9978-423f-84bc-f974dc6347b4","Type":"ContainerStarted","Data":"9c039b76d4defa4a33d53df22534cd7998d962d1489fd9ab04b73dfc51db8c37"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.342759 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kvzkk" podStartSLOduration=62.342743731 podStartE2EDuration="1m2.342743731s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.340639516 +0000 UTC m=+110.090554193" watchObservedRunningTime="2026-03-13 09:14:29.342743731 +0000 UTC m=+110.092658408" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.346518 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:29 crc kubenswrapper[4930]: E0313 09:14:29.346813 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:29.846797916 +0000 UTC m=+110.596712593 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.390716 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gc8fp" event={"ID":"635a349e-a2f9-4fe3-b58a-72334ef95d6b","Type":"ContainerStarted","Data":"0d1c91ff9aa1863e77bdd6ace633558b349856411c7ecae8822b1abe4009f2b6"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.415363 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" event={"ID":"95b6d230-5249-4b6f-a650-64de33200086","Type":"ContainerStarted","Data":"ee46200fad45b4888de1448aabca1ed9edc10f7626af42668c3d2f7411e7700f"} Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.430914 4930 ???:1] "http: TLS handshake error from 192.168.126.11:56212: no serving certificate available for the kubelet" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.431657 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk" podStartSLOduration=62.431644809 podStartE2EDuration="1m2.431644809s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.374592558 +0000 UTC m=+110.124507245" watchObservedRunningTime="2026-03-13 09:14:29.431644809 +0000 UTC m=+110.181559486" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.433031 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" podStartSLOduration=62.433023615 podStartE2EDuration="1m2.433023615s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.421958857 +0000 UTC m=+110.171873534" watchObservedRunningTime="2026-03-13 09:14:29.433023615 +0000 UTC m=+110.182938282" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.440846 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.450043 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:29 crc kubenswrapper[4930]: E0313 09:14:29.452692 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:29.952676345 +0000 UTC m=+110.702591022 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.496786 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-nk9n9" podStartSLOduration=62.496771299 podStartE2EDuration="1m2.496771299s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.448611269 +0000 UTC m=+110.198525946" watchObservedRunningTime="2026-03-13 09:14:29.496771299 +0000 UTC m=+110.246685976" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.551115 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:29 crc kubenswrapper[4930]: E0313 09:14:29.551680 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:30.051665274 +0000 UTC m=+110.801579951 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.557383 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-w957v" podStartSLOduration=63.557366682 podStartE2EDuration="1m3.557366682s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.555709229 +0000 UTC m=+110.305623906" watchObservedRunningTime="2026-03-13 09:14:29.557366682 +0000 UTC m=+110.307281359" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.557659 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-6sk6v" podStartSLOduration=62.557655 podStartE2EDuration="1m2.557655s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.520244559 +0000 UTC m=+110.270159236" watchObservedRunningTime="2026-03-13 09:14:29.557655 +0000 UTC m=+110.307569677" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.558258 4930 ???:1] "http: TLS handshake error from 192.168.126.11:56226: no serving certificate available for the kubelet" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.564031 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.591717 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podStartSLOduration=63.591700464 podStartE2EDuration="1m3.591700464s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.577523856 +0000 UTC m=+110.327438533" watchObservedRunningTime="2026-03-13 09:14:29.591700464 +0000 UTC m=+110.341615141" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.618047 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-mpbsp" podStartSLOduration=6.618029407 podStartE2EDuration="6.618029407s" podCreationTimestamp="2026-03-13 09:14:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.617803831 +0000 UTC m=+110.367718508" watchObservedRunningTime="2026-03-13 09:14:29.618029407 +0000 UTC m=+110.367944084" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.657394 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:29 crc kubenswrapper[4930]: E0313 09:14:29.657726 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:30.157713307 +0000 UTC m=+110.907627974 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.657762 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.670698 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 09:14:29 crc kubenswrapper[4930]: [-]has-synced failed: reason withheld Mar 13 09:14:29 crc kubenswrapper[4930]: [+]process-running ok Mar 13 09:14:29 crc kubenswrapper[4930]: healthz check failed Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.670760 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.674367 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" podStartSLOduration=63.674351979 podStartE2EDuration="1m3.674351979s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.673659081 +0000 UTC m=+110.423573758" watchObservedRunningTime="2026-03-13 09:14:29.674351979 +0000 UTC m=+110.424266656" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.691925 4930 ???:1] "http: TLS handshake error from 192.168.126.11:59324: no serving certificate available for the kubelet" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.763128 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:29 crc kubenswrapper[4930]: E0313 09:14:29.763477 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:30.263458762 +0000 UTC m=+111.013373439 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.773677 4930 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-nsh7c container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.14:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.773733 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" podUID="2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.14:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.776355 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podStartSLOduration=62.776336447 podStartE2EDuration="1m2.776336447s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.728099354 +0000 UTC m=+110.478014031" watchObservedRunningTime="2026-03-13 09:14:29.776336447 +0000 UTC m=+110.526251124" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.777591 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-22ldf" podStartSLOduration=63.777584479 podStartE2EDuration="1m3.777584479s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.77570866 +0000 UTC m=+110.525623337" watchObservedRunningTime="2026-03-13 09:14:29.777584479 +0000 UTC m=+110.527499156" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.814255 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" podStartSLOduration=63.814240471 podStartE2EDuration="1m3.814240471s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.813355308 +0000 UTC m=+110.563269985" watchObservedRunningTime="2026-03-13 09:14:29.814240471 +0000 UTC m=+110.564155148" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.867327 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:29 crc kubenswrapper[4930]: E0313 09:14:29.867737 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:30.367724299 +0000 UTC m=+111.117638976 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.868775 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-wrdm6" podStartSLOduration=62.868759756 podStartE2EDuration="1m2.868759756s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.842284739 +0000 UTC m=+110.592199416" watchObservedRunningTime="2026-03-13 09:14:29.868759756 +0000 UTC m=+110.618674433" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.911385 4930 ???:1] "http: TLS handshake error from 192.168.126.11:59334: no serving certificate available for the kubelet" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.942106 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-gc8fp" podStartSLOduration=62.942088769 podStartE2EDuration="1m2.942088769s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:29.900743106 +0000 UTC m=+110.650657783" watchObservedRunningTime="2026-03-13 09:14:29.942088769 +0000 UTC m=+110.692003446" Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.969853 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:29 crc kubenswrapper[4930]: E0313 09:14:29.970788 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:30.470767544 +0000 UTC m=+111.220682221 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:29 crc kubenswrapper[4930]: I0313 09:14:29.987949 4930 scope.go:117] "RemoveContainer" containerID="1fc8d89dc87d04d45f03cdf66b778aceb7d3a920f5fd254edc0f7d5491c145bf" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.071114 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:30 crc kubenswrapper[4930]: E0313 09:14:30.071395 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:30.571384426 +0000 UTC m=+111.321299093 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.173315 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:30 crc kubenswrapper[4930]: E0313 09:14:30.173898 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:30.673879275 +0000 UTC m=+111.423793952 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.290534 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:30 crc kubenswrapper[4930]: E0313 09:14:30.290838 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:30.790827001 +0000 UTC m=+111.540741678 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.313840 4930 ???:1] "http: TLS handshake error from 192.168.126.11:59338: no serving certificate available for the kubelet" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.391990 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:30 crc kubenswrapper[4930]: E0313 09:14:30.392302 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:30.892287115 +0000 UTC m=+111.642201792 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.505603 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:30 crc kubenswrapper[4930]: E0313 09:14:30.506539 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:31.00652652 +0000 UTC m=+111.756441197 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.528278 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" event={"ID":"45e7c581-937a-46b2-9501-b704239b106c","Type":"ContainerStarted","Data":"ce153486a18d08c6beb38265c56dd0ddb6bdbcbdc61e6d0391f10a7e2654ef04"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.528325 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" event={"ID":"45e7c581-937a-46b2-9501-b704239b106c","Type":"ContainerStarted","Data":"2afdd56f256bc9a5ff76720ce23ea0ff5797310242928184f64b55d39e3b89ea"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.550464 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" event={"ID":"143f048a-f720-4d97-92ac-c9766a7e39ab","Type":"ContainerStarted","Data":"e3e40cd070d4cdd05da72bed9064f18427537ff30f250f8f2e95be2d51084de2"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.551699 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.551733 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.569635 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-22ldf" event={"ID":"dc201d97-a254-45f7-8f60-f2c5dd726852","Type":"ContainerStarted","Data":"401da0a18500414bd6d51283db6b432caf9b0add3d600d776c92ee41c49a1850"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.571766 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.571795 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.608993 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:30 crc kubenswrapper[4930]: E0313 09:14:30.609389 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:31.10937425 +0000 UTC m=+111.859288927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.609573 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nptzn" event={"ID":"d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5","Type":"ContainerStarted","Data":"a2c274a8a7bc2d874e84e16482149e1d065b5730391c352b3ac0a4a722fa4340"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.616186 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rm7dk" event={"ID":"f7e799ce-dd29-42a6-9895-8ec606537229","Type":"ContainerStarted","Data":"dbb0033034a9ef38a798c926609618b956075e9deb625749a6a2bd4ee87b91dc"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.622287 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" event={"ID":"7de6cb48-405d-43d2-8387-772f05f9d85b","Type":"ContainerStarted","Data":"9f747cefca29fc1c2ce2d87192dab6f0424455387ff75df2562843a2e5fb94f6"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.636445 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" event={"ID":"bb104f34-0d73-40d9-9ba2-13401d61315d","Type":"ContainerStarted","Data":"cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.636492 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" event={"ID":"bb104f34-0d73-40d9-9ba2-13401d61315d","Type":"ContainerStarted","Data":"b59038704e56512ca177e9a750f3c1e6851fc8da6025829e26d83d6b7b07a049"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.636518 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.639246 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" event={"ID":"95528b17-57d5-4044-8c19-48dd83168399","Type":"ContainerStarted","Data":"7d2ed44dddebb293cd82be3157229530bfd2df909bced33bc0701365ffd3b8c6"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.639270 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" event={"ID":"95528b17-57d5-4044-8c19-48dd83168399","Type":"ContainerStarted","Data":"f5c4cde7b274cdd99c84f207f68f528e0b9df20bf06a21cb4264a68e6e842d7f"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.639280 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" event={"ID":"95528b17-57d5-4044-8c19-48dd83168399","Type":"ContainerStarted","Data":"c023c19b5632804dfabd3adfd08e02464f23493b0fd29102afe6e52d71068896"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.640725 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.644537 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rdbqj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.644574 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" podUID="bb104f34-0d73-40d9-9ba2-13401d61315d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.647952 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rlvdx" event={"ID":"8ea8f91d-30b3-452b-8aac-1085a5d58eab","Type":"ContainerStarted","Data":"22cb247455ea3e447128f18bc84534c01900803dea25b3737ca2cf5417b23aae"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.676604 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 09:14:30 crc kubenswrapper[4930]: [-]has-synced failed: reason withheld Mar 13 09:14:30 crc kubenswrapper[4930]: [+]process-running ok Mar 13 09:14:30 crc kubenswrapper[4930]: healthz check failed Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.676656 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.678831 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" podStartSLOduration=63.678811883 podStartE2EDuration="1m3.678811883s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:30.676903593 +0000 UTC m=+111.426818270" watchObservedRunningTime="2026-03-13 09:14:30.678811883 +0000 UTC m=+111.428726560" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.686546 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" event={"ID":"5206b83e-eee7-4ad4-bba2-e0c306b84aaf","Type":"ContainerStarted","Data":"a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.705073 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" event={"ID":"3b209dc6-6318-4983-ab60-569a294ad17f","Type":"ContainerStarted","Data":"5e8a01e8f10e42deaa4f615c0bd0a6e361492bfaf3558cfb12db24ec955e6552"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.710308 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:30 crc kubenswrapper[4930]: E0313 09:14:30.711507 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:31.211489511 +0000 UTC m=+111.961404278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.735703 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-kvzkk" event={"ID":"d8df4458-d1a4-473a-8ec3-7f8051b4cd87","Type":"ContainerStarted","Data":"1a6e3aaea3efab5415dd5baaf272426de77254e175d03b74767575d797ca0ac0"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.735724 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-xnlmh" podStartSLOduration=64.73570874 podStartE2EDuration="1m4.73570874s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:30.73495266 +0000 UTC m=+111.484867337" watchObservedRunningTime="2026-03-13 09:14:30.73570874 +0000 UTC m=+111.485623407" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.742561 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.747649 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" event={"ID":"e2e5b38e-1cee-49bf-a2c2-49919c061082","Type":"ContainerStarted","Data":"a48dd70ed1d594a5fad15bf3925f12e6bde1e850d5d919d73c712095056746b6"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.747693 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" event={"ID":"e2e5b38e-1cee-49bf-a2c2-49919c061082","Type":"ContainerStarted","Data":"643d764f6a0a8792f0955aa1d3beee93ec827fe2a4c78048a62a61d93ccfd46f"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.750019 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-w98w8" event={"ID":"514985d1-67d6-4948-b7b5-69f7dabfb424","Type":"ContainerStarted","Data":"7f0256767b2c97541e688ab1c387a20b8df68761b3521b7be93b677709ef0fef"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.750055 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-w98w8" event={"ID":"514985d1-67d6-4948-b7b5-69f7dabfb424","Type":"ContainerStarted","Data":"9b09a43d4ddae26f9b3602783c5c90be19866306f47b1bf77f774675f32bdd17"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.750752 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.751807 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-l2c8z" event={"ID":"645348bb-55ca-4eb5-abb2-845f5a021f51","Type":"ContainerStarted","Data":"99078447c76f47ea355c86037dea939d962a9c8105ec63c31d8495cdeb32ea50"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.751830 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-l2c8z" event={"ID":"645348bb-55ca-4eb5-abb2-845f5a021f51","Type":"ContainerStarted","Data":"3f98ffbaa9e7f5291bae7f529b9c84c6498820297cbb7176144f4afdb2b3450f"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.753167 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" event={"ID":"f65639e5-a600-44c3-a8b5-49aea65404e0","Type":"ContainerStarted","Data":"9a3ed9cb0a0840f8467dee30175cffc85b41d9dd0e5e058647f8d4d1745926bd"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.753191 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" event={"ID":"f65639e5-a600-44c3-a8b5-49aea65404e0","Type":"ContainerStarted","Data":"7c97db6a152cf414b16ff40c63c1199c1a20734003693cc963dfff09b6da71b3"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.777516 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-rb28t" event={"ID":"95b6d230-5249-4b6f-a650-64de33200086","Type":"ContainerStarted","Data":"fa536e58474400eafc3bf95d65bbf0a22890eab0c4c754df313dde56cea7e1a1"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.801596 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" event={"ID":"1294227b-6b51-4269-93e2-542650c9dcb8","Type":"ContainerStarted","Data":"3aec6b6b211ae40a33b555970c2e144cba59df66354694263b87f82bb67d5f63"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.814758 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.815046 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" podStartSLOduration=64.815028949 podStartE2EDuration="1m4.815028949s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:30.814871215 +0000 UTC m=+111.564785892" watchObservedRunningTime="2026-03-13 09:14:30.815028949 +0000 UTC m=+111.564943626" Mar 13 09:14:30 crc kubenswrapper[4930]: E0313 09:14:30.815784 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:31.315768788 +0000 UTC m=+112.065683465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.835801 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" event={"ID":"4f1e0b58-9978-423f-84bc-f974dc6347b4","Type":"ContainerStarted","Data":"8c960286a069fbf9bcd5f4c6337974f0f1c68a60d77a10e51960f73a826c9a7d"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.860067 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-wtfjf" event={"ID":"3f0e82b6-f4dd-4c61-88a3-31eb7f0ac724","Type":"ContainerStarted","Data":"eba94d35c135c3beb8263edb7736a8310d916ce61462fad07ceb1f8b49081683"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.872152 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" event={"ID":"d60fb09e-fa42-482c-9228-d89352573e75","Type":"ContainerStarted","Data":"d551980d1274b85a05cbf3ca612f7a73e8af7baef61f6c5b053db6dfdc5b6347"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.873059 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" podStartSLOduration=63.873037635 podStartE2EDuration="1m3.873037635s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:30.860034977 +0000 UTC m=+111.609949654" watchObservedRunningTime="2026-03-13 09:14:30.873037635 +0000 UTC m=+111.622952302" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.892667 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vfjf8" event={"ID":"8298eba9-72fc-4be8-9736-defaad400dcd","Type":"ContainerStarted","Data":"cde7c3c52d81f855971540552811e58b67d8adcfe1f4871a7208d046ed68ff83"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.920715 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-w957v" event={"ID":"a2feeb9b-9255-4660-829f-23c296a2a1c3","Type":"ContainerStarted","Data":"13aaec59d56c5086b74ff3d0f1745a7d4f6043bac526e05d5335c36cc41fe53f"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.929214 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:30 crc kubenswrapper[4930]: E0313 09:14:30.935037 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:31.435023984 +0000 UTC m=+112.184938661 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.943702 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-smklb" event={"ID":"c275e3c9-6bf5-4916-9ea5-e98baecacc87","Type":"ContainerStarted","Data":"dad3355a5bb6d7c1fc124ae917692d58349a3e23de5d2a214c372988a375ac26"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.943746 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-smklb" event={"ID":"c275e3c9-6bf5-4916-9ea5-e98baecacc87","Type":"ContainerStarted","Data":"3d0492d71899bac314b0ee3f7fa5a2c309c607b7ca710f5fd26b47b26f02b0a1"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.944648 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-w98w8" podStartSLOduration=7.9446387529999996 podStartE2EDuration="7.944638753s" podCreationTimestamp="2026-03-13 09:14:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:30.890724234 +0000 UTC m=+111.640638911" watchObservedRunningTime="2026-03-13 09:14:30.944638753 +0000 UTC m=+111.694553430" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.947180 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-mpbsp" event={"ID":"5b3d4fa3-311d-4adf-9a68-57e6a5661009","Type":"ContainerStarted","Data":"e49d954d22d93e3dfd5af9f81eb804d5f31969ab7e9de92a08b31db3e5b4a116"} Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.980059 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" podStartSLOduration=64.980043082 podStartE2EDuration="1m4.980043082s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:30.945246689 +0000 UTC m=+111.695161366" watchObservedRunningTime="2026-03-13 09:14:30.980043082 +0000 UTC m=+111.729957759" Mar 13 09:14:30 crc kubenswrapper[4930]: I0313 09:14:30.993991 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" event={"ID":"4ef93b2d-d9a0-4ef0-96d0-2568c9d825a4","Type":"ContainerStarted","Data":"92cf2f9ec0fccca8145e1e80c10ae571f6be095ddf1d68be8412e6d63a2740e2"} Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.016668 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.016829 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.022473 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.022645 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.025583 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-496vm" podStartSLOduration=64.025567334 podStartE2EDuration="1m4.025567334s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:31.022923915 +0000 UTC m=+111.772838592" watchObservedRunningTime="2026-03-13 09:14:31.025567334 +0000 UTC m=+111.775482001" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.036007 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.036360 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:31.536344614 +0000 UTC m=+112.286259291 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.056876 4930 ???:1] "http: TLS handshake error from 192.168.126.11:59352: no serving certificate available for the kubelet" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.118245 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" podStartSLOduration=64.118226759 podStartE2EDuration="1m4.118226759s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:31.076450215 +0000 UTC m=+111.826364892" watchObservedRunningTime="2026-03-13 09:14:31.118226759 +0000 UTC m=+111.868141436" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.126458 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-rlvdx" podStartSLOduration=8.126440792 podStartE2EDuration="8.126440792s" podCreationTimestamp="2026-03-13 09:14:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:31.115056356 +0000 UTC m=+111.864971033" watchObservedRunningTime="2026-03-13 09:14:31.126440792 +0000 UTC m=+111.876355459" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.147548 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.153276 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:31.653264088 +0000 UTC m=+112.403178765 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.183652 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-27rl4" podStartSLOduration=64.183630996 podStartE2EDuration="1m4.183630996s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:31.170832544 +0000 UTC m=+111.920747221" watchObservedRunningTime="2026-03-13 09:14:31.183630996 +0000 UTC m=+111.933545673" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.248398 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.248757 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:31.748739626 +0000 UTC m=+112.498654293 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.274905 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" podStartSLOduration=64.274888955 podStartE2EDuration="1m4.274888955s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:31.273870659 +0000 UTC m=+112.023785336" watchObservedRunningTime="2026-03-13 09:14:31.274888955 +0000 UTC m=+112.024803632" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.311918 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.312085 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.318035 4930 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-hjd4q container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.10:8443/livez\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.318110 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" podUID="1294227b-6b51-4269-93e2-542650c9dcb8" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.10:8443/livez\": dial tcp 10.217.0.10:8443: connect: connection refused" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.344005 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-smklb" podStartSLOduration=64.343974748 podStartE2EDuration="1m4.343974748s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:31.341811662 +0000 UTC m=+112.091726329" watchObservedRunningTime="2026-03-13 09:14:31.343974748 +0000 UTC m=+112.093889425" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.352155 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.352411 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:31.852398847 +0000 UTC m=+112.602313524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.453011 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.453380 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:31.953364848 +0000 UTC m=+112.703279515 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.554166 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.554677 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.054661877 +0000 UTC m=+112.804576554 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.655186 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.655338 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.15531095 +0000 UTC m=+112.905225627 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.655506 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.655853 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.155844604 +0000 UTC m=+112.905759281 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.657781 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 09:14:31 crc kubenswrapper[4930]: [-]has-synced failed: reason withheld Mar 13 09:14:31 crc kubenswrapper[4930]: [+]process-running ok Mar 13 09:14:31 crc kubenswrapper[4930]: healthz check failed Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.657842 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.755976 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.756188 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.256156967 +0000 UTC m=+113.006071714 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.756250 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.756591 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.256568488 +0000 UTC m=+113.006483235 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.857888 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.858088 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.358060282 +0000 UTC m=+113.107974959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.858222 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.858523 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.358513924 +0000 UTC m=+113.108428601 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.888495 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-qnw8j"] Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.921673 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.921736 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.959294 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.959487 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.459460354 +0000 UTC m=+113.209375031 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.959910 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:31 crc kubenswrapper[4930]: E0313 09:14:31.960196 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.460181963 +0000 UTC m=+113.210096630 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:31 crc kubenswrapper[4930]: I0313 09:14:31.999102 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-w98w8" event={"ID":"514985d1-67d6-4948-b7b5-69f7dabfb424","Type":"ContainerStarted","Data":"68d702e04fd9a8c77cdb8cbb332b92f9e69b4e9ae7828901c8f6f68fb6ac553e"} Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.000362 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-l2c8z" event={"ID":"645348bb-55ca-4eb5-abb2-845f5a021f51","Type":"ContainerStarted","Data":"a3034cf67f3aabf7af6a3248e834c71d681a3e484f95721166144945d8cbfb63"} Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.002373 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vfjf8" event={"ID":"8298eba9-72fc-4be8-9736-defaad400dcd","Type":"ContainerStarted","Data":"7813f1c2a24e0177d0bc60138f8c77c6c836f934445b23fa26a6f9612fb1f331"} Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.002399 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-vfjf8" event={"ID":"8298eba9-72fc-4be8-9736-defaad400dcd","Type":"ContainerStarted","Data":"17c6308523e4d7c89c0aff570d53dd06e1a65be29d2d547b7e594ea1952ceb7d"} Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.003970 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nptzn" event={"ID":"d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5","Type":"ContainerStarted","Data":"78490dbbed4601b8e73fb19760eb707d830acad9718d98f313cbc849323f74b4"} Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.004011 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-nptzn" event={"ID":"d3f426eb-df52-49eb-b7e2-be9ac0b9bfb5","Type":"ContainerStarted","Data":"4f2fdf2fe96c4a01a8bc1409a5b19470b8ab68af6ac6289fe562e956c03fd95e"} Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.005341 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" event={"ID":"4f1e0b58-9978-423f-84bc-f974dc6347b4","Type":"ContainerStarted","Data":"703c41d655cc89969ef10b93028a4fad7a9eacac34c873ac755956bfa050cae2"} Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.006783 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.007925 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"459cd20222f850789b336551de6fc7a11362c52d24e9d2ead1faf3531e81e9bd"} Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.008529 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.009932 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" event={"ID":"e2e5b38e-1cee-49bf-a2c2-49919c061082","Type":"ContainerStarted","Data":"1bd8ba79a38e24e5cbbba2599513ac0c58cf034661220007545868da6ace5bb0"} Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.011390 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.011438 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.012021 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-rdbqj container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" start-of-body= Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.012046 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" podUID="bb104f34-0d73-40d9-9ba2-13401d61315d" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.29:8080/healthz\": dial tcp 10.217.0.29:8080: connect: connection refused" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.013976 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.018809 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.024358 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.060324 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.060426 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.560411055 +0000 UTC m=+113.310325732 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.062610 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.078725 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.57870861 +0000 UTC m=+113.328623287 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.108397 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.135746 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.171498 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.171651 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.671628922 +0000 UTC m=+113.421543599 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.171976 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-l2c8z" podStartSLOduration=66.171961951 podStartE2EDuration="1m6.171961951s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:32.103348009 +0000 UTC m=+112.853262686" watchObservedRunningTime="2026-03-13 09:14:32.171961951 +0000 UTC m=+112.921876628" Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.172238 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.672227988 +0000 UTC m=+113.422142675 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.171997 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.277595 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.278028 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.778013584 +0000 UTC m=+113.527928251 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.284304 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-fl9px" podStartSLOduration=65.284287896 podStartE2EDuration="1m5.284287896s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:32.235725126 +0000 UTC m=+112.985639803" watchObservedRunningTime="2026-03-13 09:14:32.284287896 +0000 UTC m=+113.034202573" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.361682 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=26.361661185 podStartE2EDuration="26.361661185s" podCreationTimestamp="2026-03-13 09:14:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:32.338785061 +0000 UTC m=+113.088699738" watchObservedRunningTime="2026-03-13 09:14:32.361661185 +0000 UTC m=+113.111575862" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.362265 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-vfjf8" podStartSLOduration=66.362260331 podStartE2EDuration="1m6.362260331s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:32.361238684 +0000 UTC m=+113.111153361" watchObservedRunningTime="2026-03-13 09:14:32.362260331 +0000 UTC m=+113.112175008" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.381132 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.381518 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.88150377 +0000 UTC m=+113.631418447 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.407763 4930 ???:1] "http: TLS handshake error from 192.168.126.11:59358: no serving certificate available for the kubelet" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.426161 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xqlzn"] Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.426306 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-nptzn" podStartSLOduration=65.426297163 podStartE2EDuration="1m5.426297163s" podCreationTimestamp="2026-03-13 09:13:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:32.423455309 +0000 UTC m=+113.173369986" watchObservedRunningTime="2026-03-13 09:14:32.426297163 +0000 UTC m=+113.176211840" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.426398 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" podUID="1b86a263-c28c-4b03-af1e-1e57f4f203b7" containerName="controller-manager" containerID="cri-o://b6f73c03b04c6dfbedf2131baf5f9683ec7d3c8144448da982246657b98af284" gracePeriod=30 Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.484885 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.485181 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:32.985167601 +0000 UTC m=+113.735082278 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.488474 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml"] Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.587567 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.587914 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:33.087902308 +0000 UTC m=+113.837816975 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.669613 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 09:14:32 crc kubenswrapper[4930]: [-]has-synced failed: reason withheld Mar 13 09:14:32 crc kubenswrapper[4930]: [+]process-running ok Mar 13 09:14:32 crc kubenswrapper[4930]: healthz check failed Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.669664 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.672586 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tnq4n"] Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.673414 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.677836 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.680585 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tnq4n"] Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.690876 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.691152 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:33.191139378 +0000 UTC m=+113.941054055 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.772655 4930 patch_prober.go:28] interesting pod/apiserver-76f77b778f-f2kpv container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 13 09:14:32 crc kubenswrapper[4930]: [+]log ok Mar 13 09:14:32 crc kubenswrapper[4930]: [+]etcd ok Mar 13 09:14:32 crc kubenswrapper[4930]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 13 09:14:32 crc kubenswrapper[4930]: [+]poststarthook/generic-apiserver-start-informers ok Mar 13 09:14:32 crc kubenswrapper[4930]: [+]poststarthook/max-in-flight-filter ok Mar 13 09:14:32 crc kubenswrapper[4930]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 13 09:14:32 crc kubenswrapper[4930]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 13 09:14:32 crc kubenswrapper[4930]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 13 09:14:32 crc kubenswrapper[4930]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Mar 13 09:14:32 crc kubenswrapper[4930]: [+]poststarthook/project.openshift.io-projectcache ok Mar 13 09:14:32 crc kubenswrapper[4930]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 13 09:14:32 crc kubenswrapper[4930]: [-]poststarthook/openshift.io-startinformers failed: reason withheld Mar 13 09:14:32 crc kubenswrapper[4930]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 13 09:14:32 crc kubenswrapper[4930]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 13 09:14:32 crc kubenswrapper[4930]: livez check failed Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.773050 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" podUID="45e7c581-937a-46b2-9501-b704239b106c" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.792645 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/342671c3-b1f8-4c61-a0da-abe7dcdfc367-utilities\") pod \"certified-operators-tnq4n\" (UID: \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\") " pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.792703 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zqw8x\" (UniqueName: \"kubernetes.io/projected/342671c3-b1f8-4c61-a0da-abe7dcdfc367-kube-api-access-zqw8x\") pod \"certified-operators-tnq4n\" (UID: \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\") " pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.792728 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/342671c3-b1f8-4c61-a0da-abe7dcdfc367-catalog-content\") pod \"certified-operators-tnq4n\" (UID: \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\") " pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.792753 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.793013 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:33.293001682 +0000 UTC m=+114.042916359 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.827323 4930 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.857482 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-zrbn4"] Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.858325 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.860400 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.867596 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zrbn4"] Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.896640 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.896811 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:33.396785116 +0000 UTC m=+114.146699793 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.896852 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/342671c3-b1f8-4c61-a0da-abe7dcdfc367-catalog-content\") pod \"certified-operators-tnq4n\" (UID: \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\") " pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.896888 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6e93be-455f-45e4-a51b-d851591c31ed-utilities\") pod \"community-operators-zrbn4\" (UID: \"bf6e93be-455f-45e4-a51b-d851591c31ed\") " pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.896909 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7ln6\" (UniqueName: \"kubernetes.io/projected/bf6e93be-455f-45e4-a51b-d851591c31ed-kube-api-access-k7ln6\") pod \"community-operators-zrbn4\" (UID: \"bf6e93be-455f-45e4-a51b-d851591c31ed\") " pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.896945 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.896995 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6e93be-455f-45e4-a51b-d851591c31ed-catalog-content\") pod \"community-operators-zrbn4\" (UID: \"bf6e93be-455f-45e4-a51b-d851591c31ed\") " pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.897131 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/342671c3-b1f8-4c61-a0da-abe7dcdfc367-utilities\") pod \"certified-operators-tnq4n\" (UID: \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\") " pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.897204 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zqw8x\" (UniqueName: \"kubernetes.io/projected/342671c3-b1f8-4c61-a0da-abe7dcdfc367-kube-api-access-zqw8x\") pod \"certified-operators-tnq4n\" (UID: \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\") " pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.898585 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/342671c3-b1f8-4c61-a0da-abe7dcdfc367-catalog-content\") pod \"certified-operators-tnq4n\" (UID: \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\") " pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.898823 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:33.398816099 +0000 UTC m=+114.148730766 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.899159 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/342671c3-b1f8-4c61-a0da-abe7dcdfc367-utilities\") pod \"certified-operators-tnq4n\" (UID: \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\") " pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.925178 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zqw8x\" (UniqueName: \"kubernetes.io/projected/342671c3-b1f8-4c61-a0da-abe7dcdfc367-kube-api-access-zqw8x\") pod \"certified-operators-tnq4n\" (UID: \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\") " pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.997883 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.998169 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.998333 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6e93be-455f-45e4-a51b-d851591c31ed-catalog-content\") pod \"community-operators-zrbn4\" (UID: \"bf6e93be-455f-45e4-a51b-d851591c31ed\") " pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.998502 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:33.498476276 +0000 UTC m=+114.248390953 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.998612 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6e93be-455f-45e4-a51b-d851591c31ed-utilities\") pod \"community-operators-zrbn4\" (UID: \"bf6e93be-455f-45e4-a51b-d851591c31ed\") " pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.998642 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7ln6\" (UniqueName: \"kubernetes.io/projected/bf6e93be-455f-45e4-a51b-d851591c31ed-kube-api-access-k7ln6\") pod \"community-operators-zrbn4\" (UID: \"bf6e93be-455f-45e4-a51b-d851591c31ed\") " pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.998668 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.998835 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6e93be-455f-45e4-a51b-d851591c31ed-catalog-content\") pod \"community-operators-zrbn4\" (UID: \"bf6e93be-455f-45e4-a51b-d851591c31ed\") " pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:14:32 crc kubenswrapper[4930]: I0313 09:14:32.998942 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6e93be-455f-45e4-a51b-d851591c31ed-utilities\") pod \"community-operators-zrbn4\" (UID: \"bf6e93be-455f-45e4-a51b-d851591c31ed\") " pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:14:32 crc kubenswrapper[4930]: E0313 09:14:32.998952 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:33.498945078 +0000 UTC m=+114.248859745 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.020030 4930 generic.go:334] "Generic (PLEG): container finished" podID="1b86a263-c28c-4b03-af1e-1e57f4f203b7" containerID="b6f73c03b04c6dfbedf2131baf5f9683ec7d3c8144448da982246657b98af284" exitCode=0 Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.020124 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" event={"ID":"1b86a263-c28c-4b03-af1e-1e57f4f203b7","Type":"ContainerDied","Data":"b6f73c03b04c6dfbedf2131baf5f9683ec7d3c8144448da982246657b98af284"} Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.020181 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" event={"ID":"1b86a263-c28c-4b03-af1e-1e57f4f203b7","Type":"ContainerDied","Data":"dbe3e1ac734c173def970d40e3ebe38dff88c72c679d345f64762b5e8807344d"} Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.020193 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbe3e1ac734c173def970d40e3ebe38dff88c72c679d345f64762b5e8807344d" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.023403 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" event={"ID":"4f1e0b58-9978-423f-84bc-f974dc6347b4","Type":"ContainerStarted","Data":"4e38434dc61e84fdfbc5b69f7d4692030e6412a93f7dbf0acb6c8116570ea79c"} Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.023781 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" event={"ID":"4f1e0b58-9978-423f-84bc-f974dc6347b4","Type":"ContainerStarted","Data":"2b06fc9cac19434d228c566a3ca761f694dbad1e29f422c5535cbd0b7f33c935"} Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.023718 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7ln6\" (UniqueName: \"kubernetes.io/projected/bf6e93be-455f-45e4-a51b-d851591c31ed-kube-api-access-k7ln6\") pod \"community-operators-zrbn4\" (UID: \"bf6e93be-455f-45e4-a51b-d851591c31ed\") " pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.025119 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" podUID="5206b83e-eee7-4ad4-bba2-e0c306b84aaf" containerName="kube-multus-additional-cni-plugins" containerID="cri-o://a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" gracePeriod=30 Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.025945 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" podUID="04a63d8d-f300-4a8b-9ba3-5093bb7292cb" containerName="route-controller-manager" containerID="cri-o://cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600" gracePeriod=30 Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.055355 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.081664 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" podStartSLOduration=10.081644805 podStartE2EDuration="10.081644805s" podCreationTimestamp="2026-03-13 09:14:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:33.066327147 +0000 UTC m=+113.816241824" watchObservedRunningTime="2026-03-13 09:14:33.081644805 +0000 UTC m=+113.831559482" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.083655 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zp7px"] Mar 13 09:14:33 crc kubenswrapper[4930]: E0313 09:14:33.083831 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b86a263-c28c-4b03-af1e-1e57f4f203b7" containerName="controller-manager" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.083846 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b86a263-c28c-4b03-af1e-1e57f4f203b7" containerName="controller-manager" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.083929 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b86a263-c28c-4b03-af1e-1e57f4f203b7" containerName="controller-manager" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.084588 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.104675 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-client-ca\") pod \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.104751 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h7w2\" (UniqueName: \"kubernetes.io/projected/1b86a263-c28c-4b03-af1e-1e57f4f203b7-kube-api-access-9h7w2\") pod \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.104782 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-proxy-ca-bundles\") pod \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.104923 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.104946 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-config\") pod \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.104998 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b86a263-c28c-4b03-af1e-1e57f4f203b7-serving-cert\") pod \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\" (UID: \"1b86a263-c28c-4b03-af1e-1e57f4f203b7\") " Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.107610 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-client-ca" (OuterVolumeSpecName: "client-ca") pod "1b86a263-c28c-4b03-af1e-1e57f4f203b7" (UID: "1b86a263-c28c-4b03-af1e-1e57f4f203b7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.115976 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1b86a263-c28c-4b03-af1e-1e57f4f203b7" (UID: "1b86a263-c28c-4b03-af1e-1e57f4f203b7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:33 crc kubenswrapper[4930]: E0313 09:14:33.116071 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:33.616050468 +0000 UTC m=+114.365965145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.117909 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-config" (OuterVolumeSpecName: "config") pod "1b86a263-c28c-4b03-af1e-1e57f4f203b7" (UID: "1b86a263-c28c-4b03-af1e-1e57f4f203b7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.124812 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b86a263-c28c-4b03-af1e-1e57f4f203b7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1b86a263-c28c-4b03-af1e-1e57f4f203b7" (UID: "1b86a263-c28c-4b03-af1e-1e57f4f203b7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.124940 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b86a263-c28c-4b03-af1e-1e57f4f203b7-kube-api-access-9h7w2" (OuterVolumeSpecName: "kube-api-access-9h7w2") pod "1b86a263-c28c-4b03-af1e-1e57f4f203b7" (UID: "1b86a263-c28c-4b03-af1e-1e57f4f203b7"). InnerVolumeSpecName "kube-api-access-9h7w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.151416 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zp7px"] Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.174996 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.207854 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27ab43e5-6d65-4d21-a496-7c724e62b9b1-catalog-content\") pod \"certified-operators-zp7px\" (UID: \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\") " pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.207908 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.208808 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27ab43e5-6d65-4d21-a496-7c724e62b9b1-utilities\") pod \"certified-operators-zp7px\" (UID: \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\") " pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.208851 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74xgk\" (UniqueName: \"kubernetes.io/projected/27ab43e5-6d65-4d21-a496-7c724e62b9b1-kube-api-access-74xgk\") pod \"certified-operators-zp7px\" (UID: \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\") " pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.209018 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.209033 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1b86a263-c28c-4b03-af1e-1e57f4f203b7-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.209042 4930 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.209050 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h7w2\" (UniqueName: \"kubernetes.io/projected/1b86a263-c28c-4b03-af1e-1e57f4f203b7-kube-api-access-9h7w2\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.209058 4930 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1b86a263-c28c-4b03-af1e-1e57f4f203b7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:33 crc kubenswrapper[4930]: E0313 09:14:33.209307 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:33.709293418 +0000 UTC m=+114.459208095 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.275346 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-4bgrn"] Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.296715 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4bgrn"] Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.296819 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.312872 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.313206 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27ab43e5-6d65-4d21-a496-7c724e62b9b1-catalog-content\") pod \"certified-operators-zp7px\" (UID: \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\") " pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.313239 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27ab43e5-6d65-4d21-a496-7c724e62b9b1-utilities\") pod \"certified-operators-zp7px\" (UID: \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\") " pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.313260 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74xgk\" (UniqueName: \"kubernetes.io/projected/27ab43e5-6d65-4d21-a496-7c724e62b9b1-kube-api-access-74xgk\") pod \"certified-operators-zp7px\" (UID: \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\") " pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:14:33 crc kubenswrapper[4930]: E0313 09:14:33.313633 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-13 09:14:33.813618566 +0000 UTC m=+114.563533243 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.314287 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27ab43e5-6d65-4d21-a496-7c724e62b9b1-catalog-content\") pod \"certified-operators-zp7px\" (UID: \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\") " pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.314508 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27ab43e5-6d65-4d21-a496-7c724e62b9b1-utilities\") pod \"certified-operators-zp7px\" (UID: \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\") " pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.357150 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74xgk\" (UniqueName: \"kubernetes.io/projected/27ab43e5-6d65-4d21-a496-7c724e62b9b1-kube-api-access-74xgk\") pod \"certified-operators-zp7px\" (UID: \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\") " pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.415061 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d36d0eb9-1490-4b68-b637-631b7a17c10a-catalog-content\") pod \"community-operators-4bgrn\" (UID: \"d36d0eb9-1490-4b68-b637-631b7a17c10a\") " pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.415100 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5jlk\" (UniqueName: \"kubernetes.io/projected/d36d0eb9-1490-4b68-b637-631b7a17c10a-kube-api-access-g5jlk\") pod \"community-operators-4bgrn\" (UID: \"d36d0eb9-1490-4b68-b637-631b7a17c10a\") " pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.415121 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d36d0eb9-1490-4b68-b637-631b7a17c10a-utilities\") pod \"community-operators-4bgrn\" (UID: \"d36d0eb9-1490-4b68-b637-631b7a17c10a\") " pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.415147 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:33 crc kubenswrapper[4930]: E0313 09:14:33.415382 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-13 09:14:33.915369947 +0000 UTC m=+114.665284614 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-shpj8" (UID: "22f82073-0b04-4c7a-ab16-e88878214407") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.435569 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.510540 4930 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-13T09:14:32.827348684Z","Handler":null,"Name":""} Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.513102 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.516064 4930 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.516092 4930 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.524237 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.524505 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5jlk\" (UniqueName: \"kubernetes.io/projected/d36d0eb9-1490-4b68-b637-631b7a17c10a-kube-api-access-g5jlk\") pod \"community-operators-4bgrn\" (UID: \"d36d0eb9-1490-4b68-b637-631b7a17c10a\") " pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.524531 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d36d0eb9-1490-4b68-b637-631b7a17c10a-catalog-content\") pod \"community-operators-4bgrn\" (UID: \"d36d0eb9-1490-4b68-b637-631b7a17c10a\") " pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.524552 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d36d0eb9-1490-4b68-b637-631b7a17c10a-utilities\") pod \"community-operators-4bgrn\" (UID: \"d36d0eb9-1490-4b68-b637-631b7a17c10a\") " pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.524943 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d36d0eb9-1490-4b68-b637-631b7a17c10a-utilities\") pod \"community-operators-4bgrn\" (UID: \"d36d0eb9-1490-4b68-b637-631b7a17c10a\") " pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.525329 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d36d0eb9-1490-4b68-b637-631b7a17c10a-catalog-content\") pod \"community-operators-4bgrn\" (UID: \"d36d0eb9-1490-4b68-b637-631b7a17c10a\") " pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.532756 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.545256 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5jlk\" (UniqueName: \"kubernetes.io/projected/d36d0eb9-1490-4b68-b637-631b7a17c10a-kube-api-access-g5jlk\") pod \"community-operators-4bgrn\" (UID: \"d36d0eb9-1490-4b68-b637-631b7a17c10a\") " pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.570400 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tnq4n"] Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.600113 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-zrbn4"] Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.624943 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjck8\" (UniqueName: \"kubernetes.io/projected/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-kube-api-access-fjck8\") pod \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.625224 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-serving-cert\") pod \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.625264 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-client-ca\") pod \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.625286 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-config\") pod \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\" (UID: \"04a63d8d-f300-4a8b-9ba3-5093bb7292cb\") " Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.625475 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.626252 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-client-ca" (OuterVolumeSpecName: "client-ca") pod "04a63d8d-f300-4a8b-9ba3-5093bb7292cb" (UID: "04a63d8d-f300-4a8b-9ba3-5093bb7292cb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.629829 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-config" (OuterVolumeSpecName: "config") pod "04a63d8d-f300-4a8b-9ba3-5093bb7292cb" (UID: "04a63d8d-f300-4a8b-9ba3-5093bb7292cb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.630969 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "04a63d8d-f300-4a8b-9ba3-5093bb7292cb" (UID: "04a63d8d-f300-4a8b-9ba3-5093bb7292cb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.631278 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-kube-api-access-fjck8" (OuterVolumeSpecName: "kube-api-access-fjck8") pod "04a63d8d-f300-4a8b-9ba3-5093bb7292cb" (UID: "04a63d8d-f300-4a8b-9ba3-5093bb7292cb"). InnerVolumeSpecName "kube-api-access-fjck8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.632266 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.632300 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.657412 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.660912 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 09:14:33 crc kubenswrapper[4930]: [-]has-synced failed: reason withheld Mar 13 09:14:33 crc kubenswrapper[4930]: [+]process-running ok Mar 13 09:14:33 crc kubenswrapper[4930]: healthz check failed Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.660941 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.663256 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-shpj8\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.705159 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zp7px"] Mar 13 09:14:33 crc kubenswrapper[4930]: W0313 09:14:33.713571 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod27ab43e5_6d65_4d21_a496_7c724e62b9b1.slice/crio-e3d5713531526a39c7ab61b70b4dedace9d92762cb903037cda74999b6b5e858 WatchSource:0}: Error finding container e3d5713531526a39c7ab61b70b4dedace9d92762cb903037cda74999b6b5e858: Status 404 returned error can't find the container with id e3d5713531526a39c7ab61b70b4dedace9d92762cb903037cda74999b6b5e858 Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.726609 4930 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.726635 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.726646 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjck8\" (UniqueName: \"kubernetes.io/projected/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-kube-api-access-fjck8\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.726654 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/04a63d8d-f300-4a8b-9ba3-5093bb7292cb-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.874161 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-4bgrn"] Mar 13 09:14:33 crc kubenswrapper[4930]: W0313 09:14:33.884485 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd36d0eb9_1490_4b68_b637_631b7a17c10a.slice/crio-b27528e5ed93952f5748e182214de895bbdc624a57129e3b89d8d5f1a37ef85c WatchSource:0}: Error finding container b27528e5ed93952f5748e182214de895bbdc624a57129e3b89d8d5f1a37ef85c: Status 404 returned error can't find the container with id b27528e5ed93952f5748e182214de895bbdc624a57129e3b89d8d5f1a37ef85c Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.937756 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.951331 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 09:14:33 crc kubenswrapper[4930]: E0313 09:14:33.951580 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="04a63d8d-f300-4a8b-9ba3-5093bb7292cb" containerName="route-controller-manager" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.951597 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="04a63d8d-f300-4a8b-9ba3-5093bb7292cb" containerName="route-controller-manager" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.951694 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="04a63d8d-f300-4a8b-9ba3-5093bb7292cb" containerName="route-controller-manager" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.952007 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.953626 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.954059 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.959005 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 09:14:33 crc kubenswrapper[4930]: I0313 09:14:33.984487 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.026758 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.027288 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.036998 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.037319 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.037573 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69ff75e2-51e0-46e7-8a58-56b558cdde1f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"69ff75e2-51e0-46e7-8a58-56b558cdde1f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.037634 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/69ff75e2-51e0-46e7-8a58-56b558cdde1f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"69ff75e2-51e0-46e7-8a58-56b558cdde1f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.041077 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.046567 4930 generic.go:334] "Generic (PLEG): container finished" podID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" containerID="5bccd725a88473e73b456ea970fe2fc850ccb07473af0c0d31d343287c681528" exitCode=0 Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.047407 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tnq4n" event={"ID":"342671c3-b1f8-4c61-a0da-abe7dcdfc367","Type":"ContainerDied","Data":"5bccd725a88473e73b456ea970fe2fc850ccb07473af0c0d31d343287c681528"} Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.047494 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tnq4n" event={"ID":"342671c3-b1f8-4c61-a0da-abe7dcdfc367","Type":"ContainerStarted","Data":"55edb598698e63a1ac98360e709600815b6e72f8d2b5c6e8e6552c1a07631568"} Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.050155 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.052651 4930 generic.go:334] "Generic (PLEG): container finished" podID="d36d0eb9-1490-4b68-b637-631b7a17c10a" containerID="71918f546ba6672b3ec75855d8cefefc042c2bec4aae16a03fa2e37760fc83f3" exitCode=0 Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.053146 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bgrn" event={"ID":"d36d0eb9-1490-4b68-b637-631b7a17c10a","Type":"ContainerDied","Data":"71918f546ba6672b3ec75855d8cefefc042c2bec4aae16a03fa2e37760fc83f3"} Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.053193 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bgrn" event={"ID":"d36d0eb9-1490-4b68-b637-631b7a17c10a","Type":"ContainerStarted","Data":"b27528e5ed93952f5748e182214de895bbdc624a57129e3b89d8d5f1a37ef85c"} Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.062909 4930 generic.go:334] "Generic (PLEG): container finished" podID="04a63d8d-f300-4a8b-9ba3-5093bb7292cb" containerID="cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600" exitCode=0 Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.062992 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" event={"ID":"04a63d8d-f300-4a8b-9ba3-5093bb7292cb","Type":"ContainerDied","Data":"cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600"} Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.063025 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" event={"ID":"04a63d8d-f300-4a8b-9ba3-5093bb7292cb","Type":"ContainerDied","Data":"ccde838ac1d4a7709450a0bd8d9fd6947c3a0001a5607d6f258caba63ce89d40"} Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.063041 4930 scope.go:117] "RemoveContainer" containerID="cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.063061 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.064519 4930 generic.go:334] "Generic (PLEG): container finished" podID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" containerID="b0a0bac87521fa54656c88f068ede02f28d9a95a863f594ec3a60a541c7b47ec" exitCode=0 Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.064581 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zp7px" event={"ID":"27ab43e5-6d65-4d21-a496-7c724e62b9b1","Type":"ContainerDied","Data":"b0a0bac87521fa54656c88f068ede02f28d9a95a863f594ec3a60a541c7b47ec"} Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.064601 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zp7px" event={"ID":"27ab43e5-6d65-4d21-a496-7c724e62b9b1","Type":"ContainerStarted","Data":"e3d5713531526a39c7ab61b70b4dedace9d92762cb903037cda74999b6b5e858"} Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.072317 4930 generic.go:334] "Generic (PLEG): container finished" podID="bf6e93be-455f-45e4-a51b-d851591c31ed" containerID="61171bdfcce9964a938bf78cd8d5640ee622e9e15edf8c00561b6c79ba36c665" exitCode=0 Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.072644 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-xqlzn" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.073982 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrbn4" event={"ID":"bf6e93be-455f-45e4-a51b-d851591c31ed","Type":"ContainerDied","Data":"61171bdfcce9964a938bf78cd8d5640ee622e9e15edf8c00561b6c79ba36c665"} Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.074031 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrbn4" event={"ID":"bf6e93be-455f-45e4-a51b-d851591c31ed","Type":"ContainerStarted","Data":"810db0484195e739096e8ec2db9f56086c330bb79c5dbd369184adfc6fc6e9bb"} Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.114738 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml"] Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.116416 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-c5xml"] Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.139370 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/69ff75e2-51e0-46e7-8a58-56b558cdde1f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"69ff75e2-51e0-46e7-8a58-56b558cdde1f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.139539 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d33a3c8-7628-4ab8-9e9a-34624a04765e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7d33a3c8-7628-4ab8-9e9a-34624a04765e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.139580 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d33a3c8-7628-4ab8-9e9a-34624a04765e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7d33a3c8-7628-4ab8-9e9a-34624a04765e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.139694 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69ff75e2-51e0-46e7-8a58-56b558cdde1f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"69ff75e2-51e0-46e7-8a58-56b558cdde1f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.141030 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/69ff75e2-51e0-46e7-8a58-56b558cdde1f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"69ff75e2-51e0-46e7-8a58-56b558cdde1f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.153163 4930 scope.go:117] "RemoveContainer" containerID="cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.160481 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69ff75e2-51e0-46e7-8a58-56b558cdde1f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"69ff75e2-51e0-46e7-8a58-56b558cdde1f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.162194 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xqlzn"] Mar 13 09:14:34 crc kubenswrapper[4930]: E0313 09:14:34.164273 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600\": container with ID starting with cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600 not found: ID does not exist" containerID="cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.164315 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600"} err="failed to get container status \"cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600\": rpc error: code = NotFound desc = could not find container \"cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600\": container with ID starting with cf15b3099eb3ccb3b806909e92d1d8b1ffa9a42296acad60551434ebbc297600 not found: ID does not exist" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.166916 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-xqlzn"] Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.192135 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-shpj8"] Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.258069 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d33a3c8-7628-4ab8-9e9a-34624a04765e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7d33a3c8-7628-4ab8-9e9a-34624a04765e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.258126 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d33a3c8-7628-4ab8-9e9a-34624a04765e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7d33a3c8-7628-4ab8-9e9a-34624a04765e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.258256 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d33a3c8-7628-4ab8-9e9a-34624a04765e-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7d33a3c8-7628-4ab8-9e9a-34624a04765e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: W0313 09:14:34.259785 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22f82073_0b04_4c7a_ab16_e88878214407.slice/crio-b684c160c05b89dd54570071a7a23c9b05bc0b5f6ef44fe653231a7015e9f6f5 WatchSource:0}: Error finding container b684c160c05b89dd54570071a7a23c9b05bc0b5f6ef44fe653231a7015e9f6f5: Status 404 returned error can't find the container with id b684c160c05b89dd54570071a7a23c9b05bc0b5f6ef44fe653231a7015e9f6f5 Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.273238 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d33a3c8-7628-4ab8-9e9a-34624a04765e-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7d33a3c8-7628-4ab8-9e9a-34624a04765e\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.302640 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.375120 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.506351 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 13 09:14:34 crc kubenswrapper[4930]: W0313 09:14:34.510320 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod69ff75e2_51e0_46e7_8a58_56b558cdde1f.slice/crio-470f9c0641cac171bc4ea1f62722a675200f12113cca881b7f07d941fb342815 WatchSource:0}: Error finding container 470f9c0641cac171bc4ea1f62722a675200f12113cca881b7f07d941fb342815: Status 404 returned error can't find the container with id 470f9c0641cac171bc4ea1f62722a675200f12113cca881b7f07d941fb342815 Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.605415 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 13 09:14:34 crc kubenswrapper[4930]: W0313 09:14:34.630749 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod7d33a3c8_7628_4ab8_9e9a_34624a04765e.slice/crio-46d69ca330ec5a9bfd8088cd064439a9305b086d42cd7923c44b02167dd49dfe WatchSource:0}: Error finding container 46d69ca330ec5a9bfd8088cd064439a9305b086d42cd7923c44b02167dd49dfe: Status 404 returned error can't find the container with id 46d69ca330ec5a9bfd8088cd064439a9305b086d42cd7923c44b02167dd49dfe Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.659874 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 09:14:34 crc kubenswrapper[4930]: [-]has-synced failed: reason withheld Mar 13 09:14:34 crc kubenswrapper[4930]: [+]process-running ok Mar 13 09:14:34 crc kubenswrapper[4930]: healthz check failed Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.659926 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.857996 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b56p5"] Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.858937 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.861366 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.875137 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b56p5"] Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.977277 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5f9946846-zp4pk"] Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.978820 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.981194 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.981657 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.981748 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.981832 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.982019 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.982029 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.984656 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/085d31a1-d613-4e2f-a992-42bfc9b66413-utilities\") pod \"redhat-marketplace-b56p5\" (UID: \"085d31a1-d613-4e2f-a992-42bfc9b66413\") " pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.984753 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/085d31a1-d613-4e2f-a992-42bfc9b66413-catalog-content\") pod \"redhat-marketplace-b56p5\" (UID: \"085d31a1-d613-4e2f-a992-42bfc9b66413\") " pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.984803 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jl4n\" (UniqueName: \"kubernetes.io/projected/085d31a1-d613-4e2f-a992-42bfc9b66413-kube-api-access-7jl4n\") pod \"redhat-marketplace-b56p5\" (UID: \"085d31a1-d613-4e2f-a992-42bfc9b66413\") " pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.986065 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw"] Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.986820 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.991752 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.991977 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.992125 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.992264 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.992478 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.992645 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.998292 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw"] Mar 13 09:14:34 crc kubenswrapper[4930]: I0313 09:14:34.999529 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.000314 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f9946846-zp4pk"] Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.009238 4930 ???:1] "http: TLS handshake error from 192.168.126.11:59372: no serving certificate available for the kubelet" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.085770 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1efc9c9e-8858-4be8-bfec-386307b14e7d-config\") pod \"route-controller-manager-7976bf54d-zhqmw\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.085822 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/085d31a1-d613-4e2f-a992-42bfc9b66413-catalog-content\") pod \"redhat-marketplace-b56p5\" (UID: \"085d31a1-d613-4e2f-a992-42bfc9b66413\") " pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.085842 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-serving-cert\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.086326 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/085d31a1-d613-4e2f-a992-42bfc9b66413-catalog-content\") pod \"redhat-marketplace-b56p5\" (UID: \"085d31a1-d613-4e2f-a992-42bfc9b66413\") " pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.086404 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-proxy-ca-bundles\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.086471 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jl4n\" (UniqueName: \"kubernetes.io/projected/085d31a1-d613-4e2f-a992-42bfc9b66413-kube-api-access-7jl4n\") pod \"redhat-marketplace-b56p5\" (UID: \"085d31a1-d613-4e2f-a992-42bfc9b66413\") " pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.086516 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1efc9c9e-8858-4be8-bfec-386307b14e7d-client-ca\") pod \"route-controller-manager-7976bf54d-zhqmw\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.086571 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-client-ca\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.086608 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-config\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.086703 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5l7l\" (UniqueName: \"kubernetes.io/projected/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-kube-api-access-j5l7l\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.086735 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/085d31a1-d613-4e2f-a992-42bfc9b66413-utilities\") pod \"redhat-marketplace-b56p5\" (UID: \"085d31a1-d613-4e2f-a992-42bfc9b66413\") " pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.086771 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1efc9c9e-8858-4be8-bfec-386307b14e7d-serving-cert\") pod \"route-controller-manager-7976bf54d-zhqmw\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.086834 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c52h\" (UniqueName: \"kubernetes.io/projected/1efc9c9e-8858-4be8-bfec-386307b14e7d-kube-api-access-9c52h\") pod \"route-controller-manager-7976bf54d-zhqmw\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.087491 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/085d31a1-d613-4e2f-a992-42bfc9b66413-utilities\") pod \"redhat-marketplace-b56p5\" (UID: \"085d31a1-d613-4e2f-a992-42bfc9b66413\") " pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.101715 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7d33a3c8-7628-4ab8-9e9a-34624a04765e","Type":"ContainerStarted","Data":"3738ad3142fa8888f817bc69d4b3350c6bd18db20929371d821a18c86691a0eb"} Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.101784 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7d33a3c8-7628-4ab8-9e9a-34624a04765e","Type":"ContainerStarted","Data":"46d69ca330ec5a9bfd8088cd064439a9305b086d42cd7923c44b02167dd49dfe"} Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.111297 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"69ff75e2-51e0-46e7-8a58-56b558cdde1f","Type":"ContainerStarted","Data":"a3cc2ce114b4df0f5aba19e13b1d8a8ae2e506616c0e8bb900329e8e48cd16f5"} Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.111660 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"69ff75e2-51e0-46e7-8a58-56b558cdde1f","Type":"ContainerStarted","Data":"470f9c0641cac171bc4ea1f62722a675200f12113cca881b7f07d941fb342815"} Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.115626 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jl4n\" (UniqueName: \"kubernetes.io/projected/085d31a1-d613-4e2f-a992-42bfc9b66413-kube-api-access-7jl4n\") pod \"redhat-marketplace-b56p5\" (UID: \"085d31a1-d613-4e2f-a992-42bfc9b66413\") " pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.117955 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" event={"ID":"22f82073-0b04-4c7a-ab16-e88878214407","Type":"ContainerStarted","Data":"f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd"} Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.118115 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" event={"ID":"22f82073-0b04-4c7a-ab16-e88878214407","Type":"ContainerStarted","Data":"b684c160c05b89dd54570071a7a23c9b05bc0b5f6ef44fe653231a7015e9f6f5"} Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.118973 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.125660 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=1.125645382 podStartE2EDuration="1.125645382s" podCreationTimestamp="2026-03-13 09:14:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:35.12018271 +0000 UTC m=+115.870097387" watchObservedRunningTime="2026-03-13 09:14:35.125645382 +0000 UTC m=+115.875560059" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.150245 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.15022264 podStartE2EDuration="2.15022264s" podCreationTimestamp="2026-03-13 09:14:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:35.139012399 +0000 UTC m=+115.888927076" watchObservedRunningTime="2026-03-13 09:14:35.15022264 +0000 UTC m=+115.900137317" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.182777 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.187790 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1efc9c9e-8858-4be8-bfec-386307b14e7d-serving-cert\") pod \"route-controller-manager-7976bf54d-zhqmw\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.187965 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c52h\" (UniqueName: \"kubernetes.io/projected/1efc9c9e-8858-4be8-bfec-386307b14e7d-kube-api-access-9c52h\") pod \"route-controller-manager-7976bf54d-zhqmw\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.188141 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1efc9c9e-8858-4be8-bfec-386307b14e7d-config\") pod \"route-controller-manager-7976bf54d-zhqmw\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.188259 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-serving-cert\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.188389 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-proxy-ca-bundles\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.188461 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1efc9c9e-8858-4be8-bfec-386307b14e7d-client-ca\") pod \"route-controller-manager-7976bf54d-zhqmw\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.188507 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-client-ca\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.188716 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-config\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.188806 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5l7l\" (UniqueName: \"kubernetes.io/projected/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-kube-api-access-j5l7l\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.190888 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1efc9c9e-8858-4be8-bfec-386307b14e7d-config\") pod \"route-controller-manager-7976bf54d-zhqmw\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.193312 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1efc9c9e-8858-4be8-bfec-386307b14e7d-serving-cert\") pod \"route-controller-manager-7976bf54d-zhqmw\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.193322 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-config\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.193834 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-serving-cert\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.193940 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-client-ca\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.194336 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-proxy-ca-bundles\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.202230 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1efc9c9e-8858-4be8-bfec-386307b14e7d-client-ca\") pod \"route-controller-manager-7976bf54d-zhqmw\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.206845 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c52h\" (UniqueName: \"kubernetes.io/projected/1efc9c9e-8858-4be8-bfec-386307b14e7d-kube-api-access-9c52h\") pod \"route-controller-manager-7976bf54d-zhqmw\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.210535 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5l7l\" (UniqueName: \"kubernetes.io/projected/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-kube-api-access-j5l7l\") pod \"controller-manager-5f9946846-zp4pk\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.255583 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" podStartSLOduration=69.255563225 podStartE2EDuration="1m9.255563225s" podCreationTimestamp="2026-03-13 09:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:35.162343065 +0000 UTC m=+115.912257752" watchObservedRunningTime="2026-03-13 09:14:35.255563225 +0000 UTC m=+116.005477902" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.260059 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-87r56"] Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.269771 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.269788 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-87r56"] Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.309857 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.316210 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.395787 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-utilities\") pod \"redhat-marketplace-87r56\" (UID: \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\") " pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.396122 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-catalog-content\") pod \"redhat-marketplace-87r56\" (UID: \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\") " pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.396193 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frkrf\" (UniqueName: \"kubernetes.io/projected/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-kube-api-access-frkrf\") pod \"redhat-marketplace-87r56\" (UID: \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\") " pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.466059 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b56p5"] Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.497349 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frkrf\" (UniqueName: \"kubernetes.io/projected/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-kube-api-access-frkrf\") pod \"redhat-marketplace-87r56\" (UID: \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\") " pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.497737 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-utilities\") pod \"redhat-marketplace-87r56\" (UID: \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\") " pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.497782 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-catalog-content\") pod \"redhat-marketplace-87r56\" (UID: \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\") " pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.498606 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-catalog-content\") pod \"redhat-marketplace-87r56\" (UID: \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\") " pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:14:35 crc kubenswrapper[4930]: W0313 09:14:35.499448 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod085d31a1_d613_4e2f_a992_42bfc9b66413.slice/crio-0a9faadb42f5ab5c4ffdabf51f9741cf4ef9e2b5bde2eceebcb82176ded4e5cc WatchSource:0}: Error finding container 0a9faadb42f5ab5c4ffdabf51f9741cf4ef9e2b5bde2eceebcb82176ded4e5cc: Status 404 returned error can't find the container with id 0a9faadb42f5ab5c4ffdabf51f9741cf4ef9e2b5bde2eceebcb82176ded4e5cc Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.501911 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-utilities\") pod \"redhat-marketplace-87r56\" (UID: \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\") " pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.527210 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frkrf\" (UniqueName: \"kubernetes.io/projected/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-kube-api-access-frkrf\") pod \"redhat-marketplace-87r56\" (UID: \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\") " pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.609157 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.659017 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 09:14:35 crc kubenswrapper[4930]: [-]has-synced failed: reason withheld Mar 13 09:14:35 crc kubenswrapper[4930]: [+]process-running ok Mar 13 09:14:35 crc kubenswrapper[4930]: healthz check failed Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.659079 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.732499 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f9946846-zp4pk"] Mar 13 09:14:35 crc kubenswrapper[4930]: W0313 09:14:35.746728 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac6d8c7d_0e26_43fb_9508_51cddc2b5f9e.slice/crio-acd9f944ff6cf51ead8c5fe84925dc7226a9f0eaa39b95f7e39fa9ebbbcb1e49 WatchSource:0}: Error finding container acd9f944ff6cf51ead8c5fe84925dc7226a9f0eaa39b95f7e39fa9ebbbcb1e49: Status 404 returned error can't find the container with id acd9f944ff6cf51ead8c5fe84925dc7226a9f0eaa39b95f7e39fa9ebbbcb1e49 Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.837406 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-87r56"] Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.863321 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cvrjq"] Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.864252 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.865882 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 09:14:35 crc kubenswrapper[4930]: I0313 09:14:35.875343 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cvrjq"] Mar 13 09:14:35 crc kubenswrapper[4930]: W0313 09:14:35.996999 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1efc9c9e_8858_4be8_bfec_386307b14e7d.slice/crio-b68be83e4daedfd8d21fd8ec9730c5766519bc3000ddca45a5e78d1d44d34ff9 WatchSource:0}: Error finding container b68be83e4daedfd8d21fd8ec9730c5766519bc3000ddca45a5e78d1d44d34ff9: Status 404 returned error can't find the container with id b68be83e4daedfd8d21fd8ec9730c5766519bc3000ddca45a5e78d1d44d34ff9 Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.000896 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="04a63d8d-f300-4a8b-9ba3-5093bb7292cb" path="/var/lib/kubelet/pods/04a63d8d-f300-4a8b-9ba3-5093bb7292cb/volumes" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.002331 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b86a263-c28c-4b03-af1e-1e57f4f203b7" path="/var/lib/kubelet/pods/1b86a263-c28c-4b03-af1e-1e57f4f203b7/volumes" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.002896 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw"] Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.006239 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-catalog-content\") pod \"redhat-operators-cvrjq\" (UID: \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\") " pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.006351 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-utilities\") pod \"redhat-operators-cvrjq\" (UID: \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\") " pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.006494 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9bxr\" (UniqueName: \"kubernetes.io/projected/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-kube-api-access-z9bxr\") pod \"redhat-operators-cvrjq\" (UID: \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\") " pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.026644 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.041813 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.107560 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9bxr\" (UniqueName: \"kubernetes.io/projected/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-kube-api-access-z9bxr\") pod \"redhat-operators-cvrjq\" (UID: \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\") " pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.107728 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-catalog-content\") pod \"redhat-operators-cvrjq\" (UID: \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\") " pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.107777 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-utilities\") pod \"redhat-operators-cvrjq\" (UID: \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\") " pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.108483 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-utilities\") pod \"redhat-operators-cvrjq\" (UID: \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\") " pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.124810 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-catalog-content\") pod \"redhat-operators-cvrjq\" (UID: \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\") " pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.149089 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87r56" event={"ID":"6deff2ec-a8cb-4a0a-b752-bbba80f3512d","Type":"ContainerStarted","Data":"8e122a18c69e1590b7ac0e9eee2230a480d7b017cc44973dd5f44feac9390c3a"} Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.155547 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87r56" event={"ID":"6deff2ec-a8cb-4a0a-b752-bbba80f3512d","Type":"ContainerStarted","Data":"a2cce1e46af2875bab02ef1a980177811e6af1077a8e9c4530f54f8c21689c88"} Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.172640 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9bxr\" (UniqueName: \"kubernetes.io/projected/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-kube-api-access-z9bxr\") pod \"redhat-operators-cvrjq\" (UID: \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\") " pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.173657 4930 generic.go:334] "Generic (PLEG): container finished" podID="085d31a1-d613-4e2f-a992-42bfc9b66413" containerID="472c78aa0cebb2e4a504a67acbafdf7e1bfb7cf49c0202dda94c0148623caec4" exitCode=0 Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.173797 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b56p5" event={"ID":"085d31a1-d613-4e2f-a992-42bfc9b66413","Type":"ContainerDied","Data":"472c78aa0cebb2e4a504a67acbafdf7e1bfb7cf49c0202dda94c0148623caec4"} Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.173854 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b56p5" event={"ID":"085d31a1-d613-4e2f-a992-42bfc9b66413","Type":"ContainerStarted","Data":"0a9faadb42f5ab5c4ffdabf51f9741cf4ef9e2b5bde2eceebcb82176ded4e5cc"} Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.179167 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.197869 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.198717 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" event={"ID":"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e","Type":"ContainerStarted","Data":"2b72c49600ccf2bfd81c39ec804a8661af9d5bb30be58d6564f2ec08dad29736"} Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.198748 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" event={"ID":"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e","Type":"ContainerStarted","Data":"acd9f944ff6cf51ead8c5fe84925dc7226a9f0eaa39b95f7e39fa9ebbbcb1e49"} Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.199261 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.200137 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.207697 4930 patch_prober.go:28] interesting pod/console-f9d7485db-btdqm container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" start-of-body= Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.207746 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-btdqm" podUID="0805c214-0e3a-44a0-8bfe-38e8adcccde4" containerName="console" probeResult="failure" output="Get \"https://10.217.0.13:8443/health\": dial tcp 10.217.0.13:8443: connect: connection refused" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.208971 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.244283 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" podStartSLOduration=3.24426894 podStartE2EDuration="3.24426894s" podCreationTimestamp="2026-03-13 09:14:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:36.243216463 +0000 UTC m=+116.993131140" watchObservedRunningTime="2026-03-13 09:14:36.24426894 +0000 UTC m=+116.994183617" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.268095 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-78vkd"] Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.269200 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.275035 4930 generic.go:334] "Generic (PLEG): container finished" podID="7d33a3c8-7628-4ab8-9e9a-34624a04765e" containerID="3738ad3142fa8888f817bc69d4b3350c6bd18db20929371d821a18c86691a0eb" exitCode=0 Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.275098 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7d33a3c8-7628-4ab8-9e9a-34624a04765e","Type":"ContainerDied","Data":"3738ad3142fa8888f817bc69d4b3350c6bd18db20929371d821a18c86691a0eb"} Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.278542 4930 generic.go:334] "Generic (PLEG): container finished" podID="69ff75e2-51e0-46e7-8a58-56b558cdde1f" containerID="a3cc2ce114b4df0f5aba19e13b1d8a8ae2e506616c0e8bb900329e8e48cd16f5" exitCode=0 Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.278594 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"69ff75e2-51e0-46e7-8a58-56b558cdde1f","Type":"ContainerDied","Data":"a3cc2ce114b4df0f5aba19e13b1d8a8ae2e506616c0e8bb900329e8e48cd16f5"} Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.286370 4930 generic.go:334] "Generic (PLEG): container finished" podID="f65639e5-a600-44c3-a8b5-49aea65404e0" containerID="9a3ed9cb0a0840f8467dee30175cffc85b41d9dd0e5e058647f8d4d1745926bd" exitCode=0 Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.286529 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" event={"ID":"f65639e5-a600-44c3-a8b5-49aea65404e0","Type":"ContainerDied","Data":"9a3ed9cb0a0840f8467dee30175cffc85b41d9dd0e5e058647f8d4d1745926bd"} Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.289191 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" event={"ID":"1efc9c9e-8858-4be8-bfec-386307b14e7d","Type":"ContainerStarted","Data":"b68be83e4daedfd8d21fd8ec9730c5766519bc3000ddca45a5e78d1d44d34ff9"} Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.291935 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-78vkd"] Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.338105 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.344749 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.415118 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/980ffd8b-d762-4aac-8037-ccab15a76309-utilities\") pod \"redhat-operators-78vkd\" (UID: \"980ffd8b-d762-4aac-8037-ccab15a76309\") " pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.415426 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/980ffd8b-d762-4aac-8037-ccab15a76309-catalog-content\") pod \"redhat-operators-78vkd\" (UID: \"980ffd8b-d762-4aac-8037-ccab15a76309\") " pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.415466 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvxdj\" (UniqueName: \"kubernetes.io/projected/980ffd8b-d762-4aac-8037-ccab15a76309-kube-api-access-rvxdj\") pod \"redhat-operators-78vkd\" (UID: \"980ffd8b-d762-4aac-8037-ccab15a76309\") " pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.516600 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/980ffd8b-d762-4aac-8037-ccab15a76309-utilities\") pod \"redhat-operators-78vkd\" (UID: \"980ffd8b-d762-4aac-8037-ccab15a76309\") " pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.516648 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/980ffd8b-d762-4aac-8037-ccab15a76309-catalog-content\") pod \"redhat-operators-78vkd\" (UID: \"980ffd8b-d762-4aac-8037-ccab15a76309\") " pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.516684 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvxdj\" (UniqueName: \"kubernetes.io/projected/980ffd8b-d762-4aac-8037-ccab15a76309-kube-api-access-rvxdj\") pod \"redhat-operators-78vkd\" (UID: \"980ffd8b-d762-4aac-8037-ccab15a76309\") " pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.517071 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/980ffd8b-d762-4aac-8037-ccab15a76309-utilities\") pod \"redhat-operators-78vkd\" (UID: \"980ffd8b-d762-4aac-8037-ccab15a76309\") " pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.517357 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/980ffd8b-d762-4aac-8037-ccab15a76309-catalog-content\") pod \"redhat-operators-78vkd\" (UID: \"980ffd8b-d762-4aac-8037-ccab15a76309\") " pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.558590 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvxdj\" (UniqueName: \"kubernetes.io/projected/980ffd8b-d762-4aac-8037-ccab15a76309-kube-api-access-rvxdj\") pod \"redhat-operators-78vkd\" (UID: \"980ffd8b-d762-4aac-8037-ccab15a76309\") " pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.629138 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cvrjq"] Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.640254 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.655575 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.667824 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 09:14:36 crc kubenswrapper[4930]: [-]has-synced failed: reason withheld Mar 13 09:14:36 crc kubenswrapper[4930]: [+]process-running ok Mar 13 09:14:36 crc kubenswrapper[4930]: healthz check failed Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.667879 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.720298 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.720356 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.720454 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.720505 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 09:14:36 crc kubenswrapper[4930]: E0313 09:14:36.900965 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 13 09:14:36 crc kubenswrapper[4930]: E0313 09:14:36.907551 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 13 09:14:36 crc kubenswrapper[4930]: E0313 09:14:36.941148 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 13 09:14:36 crc kubenswrapper[4930]: E0313 09:14:36.941200 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" podUID="5206b83e-eee7-4ad4-bba2-e0c306b84aaf" containerName="kube-multus-additional-cni-plugins" Mar 13 09:14:36 crc kubenswrapper[4930]: I0313 09:14:36.980951 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.026686 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.027843 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.058505 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.084340 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=1.084322816 podStartE2EDuration="1.084322816s" podCreationTimestamp="2026-03-13 09:14:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:37.079247995 +0000 UTC m=+117.829162672" watchObservedRunningTime="2026-03-13 09:14:37.084322816 +0000 UTC m=+117.834237493" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.131648 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.131747 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.131819 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.139108 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.139160 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.141121 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.190381 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-78vkd"] Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.314682 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.315712 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78vkd" event={"ID":"980ffd8b-d762-4aac-8037-ccab15a76309","Type":"ContainerStarted","Data":"536da4b970ca3ac794fb8d2236f2ab9e9197fa93e1ce5e4e876024bdbe57c2fa"} Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.321303 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" event={"ID":"1efc9c9e-8858-4be8-bfec-386307b14e7d","Type":"ContainerStarted","Data":"e983976c294bc96dd5a09833a6ccdd77e91b0b72c6d0e46244729872e40ec29c"} Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.321490 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.341918 4930 generic.go:334] "Generic (PLEG): container finished" podID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" containerID="8e122a18c69e1590b7ac0e9eee2230a480d7b017cc44973dd5f44feac9390c3a" exitCode=0 Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.343508 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.344686 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87r56" event={"ID":"6deff2ec-a8cb-4a0a-b752-bbba80f3512d","Type":"ContainerDied","Data":"8e122a18c69e1590b7ac0e9eee2230a480d7b017cc44973dd5f44feac9390c3a"} Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.344885 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.345611 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" podStartSLOduration=4.345594348 podStartE2EDuration="4.345594348s" podCreationTimestamp="2026-03-13 09:14:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:14:37.342647501 +0000 UTC m=+118.092562178" watchObservedRunningTime="2026-03-13 09:14:37.345594348 +0000 UTC m=+118.095509025" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.355245 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.366344 4930 generic.go:334] "Generic (PLEG): container finished" podID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerID="7ee210e26eaadde01e83f39b637dc5c600bd87d87402fcc3f3a29904b1b03c34" exitCode=0 Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.367152 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvrjq" event={"ID":"5933708c-4e7f-4567-bd1e-df3dfe85fcd5","Type":"ContainerDied","Data":"7ee210e26eaadde01e83f39b637dc5c600bd87d87402fcc3f3a29904b1b03c34"} Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.367182 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvrjq" event={"ID":"5933708c-4e7f-4567-bd1e-df3dfe85fcd5","Type":"ContainerStarted","Data":"3b718bc407c8d2708b66369e8b498b225b48037371e4cc333f9029d9cc83be8a"} Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.660791 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.667500 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.688956 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.744506 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d33a3c8-7628-4ab8-9e9a-34624a04765e-kubelet-dir\") pod \"7d33a3c8-7628-4ab8-9e9a-34624a04765e\" (UID: \"7d33a3c8-7628-4ab8-9e9a-34624a04765e\") " Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.744591 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d33a3c8-7628-4ab8-9e9a-34624a04765e-kube-api-access\") pod \"7d33a3c8-7628-4ab8-9e9a-34624a04765e\" (UID: \"7d33a3c8-7628-4ab8-9e9a-34624a04765e\") " Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.744647 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7d33a3c8-7628-4ab8-9e9a-34624a04765e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7d33a3c8-7628-4ab8-9e9a-34624a04765e" (UID: "7d33a3c8-7628-4ab8-9e9a-34624a04765e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.744930 4930 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7d33a3c8-7628-4ab8-9e9a-34624a04765e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.754157 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d33a3c8-7628-4ab8-9e9a-34624a04765e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7d33a3c8-7628-4ab8-9e9a-34624a04765e" (UID: "7d33a3c8-7628-4ab8-9e9a-34624a04765e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.834734 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.843557 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.847103 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d33a3c8-7628-4ab8-9e9a-34624a04765e-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.948024 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f65639e5-a600-44c3-a8b5-49aea65404e0-config-volume\") pod \"f65639e5-a600-44c3-a8b5-49aea65404e0\" (UID: \"f65639e5-a600-44c3-a8b5-49aea65404e0\") " Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.948077 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f65639e5-a600-44c3-a8b5-49aea65404e0-secret-volume\") pod \"f65639e5-a600-44c3-a8b5-49aea65404e0\" (UID: \"f65639e5-a600-44c3-a8b5-49aea65404e0\") " Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.948116 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69ff75e2-51e0-46e7-8a58-56b558cdde1f-kube-api-access\") pod \"69ff75e2-51e0-46e7-8a58-56b558cdde1f\" (UID: \"69ff75e2-51e0-46e7-8a58-56b558cdde1f\") " Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.948148 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/69ff75e2-51e0-46e7-8a58-56b558cdde1f-kubelet-dir\") pod \"69ff75e2-51e0-46e7-8a58-56b558cdde1f\" (UID: \"69ff75e2-51e0-46e7-8a58-56b558cdde1f\") " Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.948279 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nwvg\" (UniqueName: \"kubernetes.io/projected/f65639e5-a600-44c3-a8b5-49aea65404e0-kube-api-access-4nwvg\") pod \"f65639e5-a600-44c3-a8b5-49aea65404e0\" (UID: \"f65639e5-a600-44c3-a8b5-49aea65404e0\") " Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.949723 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/69ff75e2-51e0-46e7-8a58-56b558cdde1f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "69ff75e2-51e0-46e7-8a58-56b558cdde1f" (UID: "69ff75e2-51e0-46e7-8a58-56b558cdde1f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.950659 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f65639e5-a600-44c3-a8b5-49aea65404e0-config-volume" (OuterVolumeSpecName: "config-volume") pod "f65639e5-a600-44c3-a8b5-49aea65404e0" (UID: "f65639e5-a600-44c3-a8b5-49aea65404e0"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.954503 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f65639e5-a600-44c3-a8b5-49aea65404e0-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f65639e5-a600-44c3-a8b5-49aea65404e0" (UID: "f65639e5-a600-44c3-a8b5-49aea65404e0"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.954953 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69ff75e2-51e0-46e7-8a58-56b558cdde1f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "69ff75e2-51e0-46e7-8a58-56b558cdde1f" (UID: "69ff75e2-51e0-46e7-8a58-56b558cdde1f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:37 crc kubenswrapper[4930]: I0313 09:14:37.955001 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f65639e5-a600-44c3-a8b5-49aea65404e0-kube-api-access-4nwvg" (OuterVolumeSpecName: "kube-api-access-4nwvg") pod "f65639e5-a600-44c3-a8b5-49aea65404e0" (UID: "f65639e5-a600-44c3-a8b5-49aea65404e0"). InnerVolumeSpecName "kube-api-access-4nwvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.054083 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nwvg\" (UniqueName: \"kubernetes.io/projected/f65639e5-a600-44c3-a8b5-49aea65404e0-kube-api-access-4nwvg\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.054115 4930 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f65639e5-a600-44c3-a8b5-49aea65404e0-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.054124 4930 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f65639e5-a600-44c3-a8b5-49aea65404e0-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.054133 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/69ff75e2-51e0-46e7-8a58-56b558cdde1f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.054143 4930 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/69ff75e2-51e0-46e7-8a58-56b558cdde1f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 09:14:38 crc kubenswrapper[4930]: W0313 09:14:38.089570 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-74279bac7958e117806bf0b3be883b4637521f4544664f9b120c9f8a573a6e7c WatchSource:0}: Error finding container 74279bac7958e117806bf0b3be883b4637521f4544664f9b120c9f8a573a6e7c: Status 404 returned error can't find the container with id 74279bac7958e117806bf0b3be883b4637521f4544664f9b120c9f8a573a6e7c Mar 13 09:14:38 crc kubenswrapper[4930]: W0313 09:14:38.094493 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-052fff58177eb591211f45e0ccc215445fe9bd54f8a6e39905b76ab776302f4f WatchSource:0}: Error finding container 052fff58177eb591211f45e0ccc215445fe9bd54f8a6e39905b76ab776302f4f: Status 404 returned error can't find the container with id 052fff58177eb591211f45e0ccc215445fe9bd54f8a6e39905b76ab776302f4f Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.380736 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.380801 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"69ff75e2-51e0-46e7-8a58-56b558cdde1f","Type":"ContainerDied","Data":"470f9c0641cac171bc4ea1f62722a675200f12113cca881b7f07d941fb342815"} Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.380971 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="470f9c0641cac171bc4ea1f62722a675200f12113cca881b7f07d941fb342815" Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.385415 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" event={"ID":"f65639e5-a600-44c3-a8b5-49aea65404e0","Type":"ContainerDied","Data":"7c97db6a152cf414b16ff40c63c1199c1a20734003693cc963dfff09b6da71b3"} Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.385490 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c97db6a152cf414b16ff40c63c1199c1a20734003693cc963dfff09b6da71b3" Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.385716 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz" Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.393939 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"052fff58177eb591211f45e0ccc215445fe9bd54f8a6e39905b76ab776302f4f"} Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.404063 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"74279bac7958e117806bf0b3be883b4637521f4544664f9b120c9f8a573a6e7c"} Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.409720 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"bb58796f982dcfd44328ca5847b5bc792a6a82a841057b892a0ccf791dc159ed"} Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.409747 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"2c9e873389cc432249910f82aff362f32a4590baec99b02f32e57c2281748642"} Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.409973 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.418525 4930 generic.go:334] "Generic (PLEG): container finished" podID="980ffd8b-d762-4aac-8037-ccab15a76309" containerID="84c1e2558aa40b4e4217c5eab421123a53239ae0ed0be2363278282d1521b81e" exitCode=0 Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.418596 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78vkd" event={"ID":"980ffd8b-d762-4aac-8037-ccab15a76309","Type":"ContainerDied","Data":"84c1e2558aa40b4e4217c5eab421123a53239ae0ed0be2363278282d1521b81e"} Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.425301 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7d33a3c8-7628-4ab8-9e9a-34624a04765e","Type":"ContainerDied","Data":"46d69ca330ec5a9bfd8088cd064439a9305b086d42cd7923c44b02167dd49dfe"} Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.425621 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46d69ca330ec5a9bfd8088cd064439a9305b086d42cd7923c44b02167dd49dfe" Mar 13 09:14:38 crc kubenswrapper[4930]: I0313 09:14:38.426054 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 13 09:14:39 crc kubenswrapper[4930]: I0313 09:14:39.176675 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-w98w8" Mar 13 09:14:39 crc kubenswrapper[4930]: I0313 09:14:39.486800 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"26807ae97c9f6a2dc11178252a5713b39805216fdfc63ec65de5dd78037a67c7"} Mar 13 09:14:39 crc kubenswrapper[4930]: I0313 09:14:39.494779 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a097179c3b80383929648aaed3a74d8fba189211b78f8d2bb5a2e6716bce0db2"} Mar 13 09:14:39 crc kubenswrapper[4930]: I0313 09:14:39.832218 4930 ???:1] "http: TLS handshake error from 192.168.126.11:41616: no serving certificate available for the kubelet" Mar 13 09:14:40 crc kubenswrapper[4930]: I0313 09:14:40.159547 4930 ???:1] "http: TLS handshake error from 192.168.126.11:41626: no serving certificate available for the kubelet" Mar 13 09:14:42 crc kubenswrapper[4930]: I0313 09:14:42.439090 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:14:46 crc kubenswrapper[4930]: I0313 09:14:46.213136 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:46 crc kubenswrapper[4930]: I0313 09:14:46.216929 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:14:46 crc kubenswrapper[4930]: I0313 09:14:46.723630 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 09:14:46 crc kubenswrapper[4930]: I0313 09:14:46.810164 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:14:46 crc kubenswrapper[4930]: E0313 09:14:46.902115 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 13 09:14:46 crc kubenswrapper[4930]: E0313 09:14:46.903720 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 13 09:14:46 crc kubenswrapper[4930]: E0313 09:14:46.905059 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 13 09:14:46 crc kubenswrapper[4930]: E0313 09:14:46.905090 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" podUID="5206b83e-eee7-4ad4-bba2-e0c306b84aaf" containerName="kube-multus-additional-cni-plugins" Mar 13 09:14:51 crc kubenswrapper[4930]: I0313 09:14:51.951188 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5f9946846-zp4pk"] Mar 13 09:14:51 crc kubenswrapper[4930]: I0313 09:14:51.952542 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" podUID="ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" containerName="controller-manager" containerID="cri-o://2b72c49600ccf2bfd81c39ec804a8661af9d5bb30be58d6564f2ec08dad29736" gracePeriod=30 Mar 13 09:14:51 crc kubenswrapper[4930]: I0313 09:14:51.967241 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw"] Mar 13 09:14:51 crc kubenswrapper[4930]: I0313 09:14:51.967560 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" podUID="1efc9c9e-8858-4be8-bfec-386307b14e7d" containerName="route-controller-manager" containerID="cri-o://e983976c294bc96dd5a09833a6ccdd77e91b0b72c6d0e46244729872e40ec29c" gracePeriod=30 Mar 13 09:14:52 crc kubenswrapper[4930]: I0313 09:14:52.642476 4930 generic.go:334] "Generic (PLEG): container finished" podID="ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" containerID="2b72c49600ccf2bfd81c39ec804a8661af9d5bb30be58d6564f2ec08dad29736" exitCode=0 Mar 13 09:14:52 crc kubenswrapper[4930]: I0313 09:14:52.642549 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" event={"ID":"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e","Type":"ContainerDied","Data":"2b72c49600ccf2bfd81c39ec804a8661af9d5bb30be58d6564f2ec08dad29736"} Mar 13 09:14:53 crc kubenswrapper[4930]: I0313 09:14:53.648930 4930 generic.go:334] "Generic (PLEG): container finished" podID="1efc9c9e-8858-4be8-bfec-386307b14e7d" containerID="e983976c294bc96dd5a09833a6ccdd77e91b0b72c6d0e46244729872e40ec29c" exitCode=0 Mar 13 09:14:53 crc kubenswrapper[4930]: I0313 09:14:53.648982 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" event={"ID":"1efc9c9e-8858-4be8-bfec-386307b14e7d","Type":"ContainerDied","Data":"e983976c294bc96dd5a09833a6ccdd77e91b0b72c6d0e46244729872e40ec29c"} Mar 13 09:14:53 crc kubenswrapper[4930]: I0313 09:14:53.947734 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:14:56 crc kubenswrapper[4930]: I0313 09:14:56.311309 4930 patch_prober.go:28] interesting pod/controller-manager-5f9946846-zp4pk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 09:14:56 crc kubenswrapper[4930]: I0313 09:14:56.311736 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" podUID="ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 09:14:56 crc kubenswrapper[4930]: I0313 09:14:56.317720 4930 patch_prober.go:28] interesting pod/route-controller-manager-7976bf54d-zhqmw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.52:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 09:14:56 crc kubenswrapper[4930]: I0313 09:14:56.317772 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" podUID="1efc9c9e-8858-4be8-bfec-386307b14e7d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.52:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 09:14:56 crc kubenswrapper[4930]: E0313 09:14:56.900794 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 13 09:14:56 crc kubenswrapper[4930]: E0313 09:14:56.902829 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 13 09:14:56 crc kubenswrapper[4930]: E0313 09:14:56.903781 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 13 09:14:56 crc kubenswrapper[4930]: E0313 09:14:56.903816 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" podUID="5206b83e-eee7-4ad4-bba2-e0c306b84aaf" containerName="kube-multus-additional-cni-plugins" Mar 13 09:14:56 crc kubenswrapper[4930]: I0313 09:14:56.991424 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.009850 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=4.009831471 podStartE2EDuration="4.009831471s" podCreationTimestamp="2026-03-13 09:14:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:15:00.007735884 +0000 UTC m=+140.757650561" watchObservedRunningTime="2026-03-13 09:15:00.009831471 +0000 UTC m=+140.759746148" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.144999 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b"] Mar 13 09:15:00 crc kubenswrapper[4930]: E0313 09:15:00.147890 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f65639e5-a600-44c3-a8b5-49aea65404e0" containerName="collect-profiles" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.147916 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f65639e5-a600-44c3-a8b5-49aea65404e0" containerName="collect-profiles" Mar 13 09:15:00 crc kubenswrapper[4930]: E0313 09:15:00.147929 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69ff75e2-51e0-46e7-8a58-56b558cdde1f" containerName="pruner" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.147939 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="69ff75e2-51e0-46e7-8a58-56b558cdde1f" containerName="pruner" Mar 13 09:15:00 crc kubenswrapper[4930]: E0313 09:15:00.147952 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d33a3c8-7628-4ab8-9e9a-34624a04765e" containerName="pruner" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.147962 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d33a3c8-7628-4ab8-9e9a-34624a04765e" containerName="pruner" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.148173 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f65639e5-a600-44c3-a8b5-49aea65404e0" containerName="collect-profiles" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.148192 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="69ff75e2-51e0-46e7-8a58-56b558cdde1f" containerName="pruner" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.148210 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d33a3c8-7628-4ab8-9e9a-34624a04765e" containerName="pruner" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.148732 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.155563 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b"] Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.209168 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.209251 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.209266 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p8k95\" (UniqueName: \"kubernetes.io/projected/58844174-f3d8-4618-a396-78d721d893b8-kube-api-access-p8k95\") pod \"collect-profiles-29556555-svg2b\" (UID: \"58844174-f3d8-4618-a396-78d721d893b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.209311 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58844174-f3d8-4618-a396-78d721d893b8-config-volume\") pod \"collect-profiles-29556555-svg2b\" (UID: \"58844174-f3d8-4618-a396-78d721d893b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.209407 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58844174-f3d8-4618-a396-78d721d893b8-secret-volume\") pod \"collect-profiles-29556555-svg2b\" (UID: \"58844174-f3d8-4618-a396-78d721d893b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.310962 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p8k95\" (UniqueName: \"kubernetes.io/projected/58844174-f3d8-4618-a396-78d721d893b8-kube-api-access-p8k95\") pod \"collect-profiles-29556555-svg2b\" (UID: \"58844174-f3d8-4618-a396-78d721d893b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.311038 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58844174-f3d8-4618-a396-78d721d893b8-config-volume\") pod \"collect-profiles-29556555-svg2b\" (UID: \"58844174-f3d8-4618-a396-78d721d893b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.311258 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58844174-f3d8-4618-a396-78d721d893b8-secret-volume\") pod \"collect-profiles-29556555-svg2b\" (UID: \"58844174-f3d8-4618-a396-78d721d893b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.312198 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58844174-f3d8-4618-a396-78d721d893b8-config-volume\") pod \"collect-profiles-29556555-svg2b\" (UID: \"58844174-f3d8-4618-a396-78d721d893b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.322088 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58844174-f3d8-4618-a396-78d721d893b8-secret-volume\") pod \"collect-profiles-29556555-svg2b\" (UID: \"58844174-f3d8-4618-a396-78d721d893b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.325760 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p8k95\" (UniqueName: \"kubernetes.io/projected/58844174-f3d8-4618-a396-78d721d893b8-kube-api-access-p8k95\") pod \"collect-profiles-29556555-svg2b\" (UID: \"58844174-f3d8-4618-a396-78d721d893b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.530359 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:00 crc kubenswrapper[4930]: I0313 09:15:00.666728 4930 ???:1] "http: TLS handshake error from 192.168.126.11:47176: no serving certificate available for the kubelet" Mar 13 09:15:03 crc kubenswrapper[4930]: E0313 09:15:03.536187 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 13 09:15:03 crc kubenswrapper[4930]: E0313 09:15:03.536707 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-z9bxr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-cvrjq_openshift-marketplace(5933708c-4e7f-4567-bd1e-df3dfe85fcd5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 09:15:03 crc kubenswrapper[4930]: E0313 09:15:03.537971 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-cvrjq" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" Mar 13 09:15:03 crc kubenswrapper[4930]: I0313 09:15:03.715323 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-qnw8j_5206b83e-eee7-4ad4-bba2-e0c306b84aaf/kube-multus-additional-cni-plugins/0.log" Mar 13 09:15:03 crc kubenswrapper[4930]: I0313 09:15:03.715407 4930 generic.go:334] "Generic (PLEG): container finished" podID="5206b83e-eee7-4ad4-bba2-e0c306b84aaf" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" exitCode=137 Mar 13 09:15:03 crc kubenswrapper[4930]: I0313 09:15:03.715525 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" event={"ID":"5206b83e-eee7-4ad4-bba2-e0c306b84aaf","Type":"ContainerDied","Data":"a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486"} Mar 13 09:15:06 crc kubenswrapper[4930]: E0313 09:15:06.011777 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-cvrjq" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" Mar 13 09:15:06 crc kubenswrapper[4930]: I0313 09:15:06.311239 4930 patch_prober.go:28] interesting pod/controller-manager-5f9946846-zp4pk container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.51:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 09:15:06 crc kubenswrapper[4930]: I0313 09:15:06.311584 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" podUID="ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.51:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 09:15:06 crc kubenswrapper[4930]: I0313 09:15:06.317484 4930 patch_prober.go:28] interesting pod/route-controller-manager-7976bf54d-zhqmw container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.52:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 09:15:06 crc kubenswrapper[4930]: I0313 09:15:06.318117 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" podUID="1efc9c9e-8858-4be8-bfec-386307b14e7d" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.52:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 09:15:06 crc kubenswrapper[4930]: E0313 09:15:06.440671 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Mar 13 09:15:06 crc kubenswrapper[4930]: E0313 09:15:06.440805 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rvxdj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-78vkd_openshift-marketplace(980ffd8b-d762-4aac-8037-ccab15a76309): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 09:15:06 crc kubenswrapper[4930]: E0313 09:15:06.442829 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-78vkd" podUID="980ffd8b-d762-4aac-8037-ccab15a76309" Mar 13 09:15:06 crc kubenswrapper[4930]: E0313 09:15:06.528560 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 13 09:15:06 crc kubenswrapper[4930]: E0313 09:15:06.528707 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g5jlk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-4bgrn_openshift-marketplace(d36d0eb9-1490-4b68-b637-631b7a17c10a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 09:15:06 crc kubenswrapper[4930]: E0313 09:15:06.529935 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-4bgrn" podUID="d36d0eb9-1490-4b68-b637-631b7a17c10a" Mar 13 09:15:06 crc kubenswrapper[4930]: E0313 09:15:06.899295 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486 is running failed: container process not found" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 13 09:15:06 crc kubenswrapper[4930]: E0313 09:15:06.899787 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486 is running failed: container process not found" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 13 09:15:06 crc kubenswrapper[4930]: E0313 09:15:06.900275 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486 is running failed: container process not found" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" cmd=["/bin/bash","-c","test -f /ready/ready"] Mar 13 09:15:06 crc kubenswrapper[4930]: E0313 09:15:06.900313 4930 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486 is running failed: container process not found" probeType="Readiness" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" podUID="5206b83e-eee7-4ad4-bba2-e0c306b84aaf" containerName="kube-multus-additional-cni-plugins" Mar 13 09:15:07 crc kubenswrapper[4930]: I0313 09:15:07.137906 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" Mar 13 09:15:07 crc kubenswrapper[4930]: E0313 09:15:07.584408 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-4bgrn" podUID="d36d0eb9-1490-4b68-b637-631b7a17c10a" Mar 13 09:15:07 crc kubenswrapper[4930]: E0313 09:15:07.585053 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-78vkd" podUID="980ffd8b-d762-4aac-8037-ccab15a76309" Mar 13 09:15:07 crc kubenswrapper[4930]: E0313 09:15:07.651314 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 13 09:15:07 crc kubenswrapper[4930]: E0313 09:15:07.651497 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7jl4n,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-b56p5_openshift-marketplace(085d31a1-d613-4e2f-a992-42bfc9b66413): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 09:15:07 crc kubenswrapper[4930]: E0313 09:15:07.652980 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-b56p5" podUID="085d31a1-d613-4e2f-a992-42bfc9b66413" Mar 13 09:15:07 crc kubenswrapper[4930]: E0313 09:15:07.674171 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Mar 13 09:15:07 crc kubenswrapper[4930]: E0313 09:15:07.674378 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-frkrf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-87r56_openshift-marketplace(6deff2ec-a8cb-4a0a-b752-bbba80f3512d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 09:15:07 crc kubenswrapper[4930]: E0313 09:15:07.675850 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-87r56" podUID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" Mar 13 09:15:08 crc kubenswrapper[4930]: I0313 09:15:08.827244 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 09:15:08 crc kubenswrapper[4930]: I0313 09:15:08.827848 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 09:15:08 crc kubenswrapper[4930]: I0313 09:15:08.835589 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 13 09:15:08 crc kubenswrapper[4930]: I0313 09:15:08.835688 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 13 09:15:08 crc kubenswrapper[4930]: I0313 09:15:08.837735 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 09:15:08 crc kubenswrapper[4930]: I0313 09:15:08.930483 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/74eb4334-fb07-42c4-b59f-7a07546af2e9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"74eb4334-fb07-42c4-b59f-7a07546af2e9\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 09:15:08 crc kubenswrapper[4930]: I0313 09:15:08.930760 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74eb4334-fb07-42c4-b59f-7a07546af2e9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"74eb4334-fb07-42c4-b59f-7a07546af2e9\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.032115 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74eb4334-fb07-42c4-b59f-7a07546af2e9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"74eb4334-fb07-42c4-b59f-7a07546af2e9\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.032192 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/74eb4334-fb07-42c4-b59f-7a07546af2e9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"74eb4334-fb07-42c4-b59f-7a07546af2e9\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.032260 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/74eb4334-fb07-42c4-b59f-7a07546af2e9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"74eb4334-fb07-42c4-b59f-7a07546af2e9\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.049816 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74eb4334-fb07-42c4-b59f-7a07546af2e9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"74eb4334-fb07-42c4-b59f-7a07546af2e9\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.151214 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.303400 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-b56p5" podUID="085d31a1-d613-4e2f-a992-42bfc9b66413" Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.308484 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-87r56" podUID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.380215 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.380604 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-74xgk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-zp7px_openshift-marketplace(27ab43e5-6d65-4d21-a496-7c724e62b9b1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.381752 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-zp7px" podUID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.397212 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.405395 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.412093 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.412188 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zqw8x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-tnq4n_openshift-marketplace(342671c3-b1f8-4c61-a0da-abe7dcdfc367): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.413289 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-tnq4n" podUID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.414961 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-qnw8j_5206b83e-eee7-4ad4-bba2-e0c306b84aaf/kube-multus-additional-cni-plugins/0.log" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.415016 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.426453 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-59795d48dc-ndw77"] Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.426672 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5206b83e-eee7-4ad4-bba2-e0c306b84aaf" containerName="kube-multus-additional-cni-plugins" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.426684 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="5206b83e-eee7-4ad4-bba2-e0c306b84aaf" containerName="kube-multus-additional-cni-plugins" Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.426704 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1efc9c9e-8858-4be8-bfec-386307b14e7d" containerName="route-controller-manager" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.426714 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="1efc9c9e-8858-4be8-bfec-386307b14e7d" containerName="route-controller-manager" Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.426730 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" containerName="controller-manager" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.426737 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" containerName="controller-manager" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.442884 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" containerName="controller-manager" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.444588 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="1efc9c9e-8858-4be8-bfec-386307b14e7d" containerName="route-controller-manager" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.444623 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="5206b83e-eee7-4ad4-bba2-e0c306b84aaf" containerName="kube-multus-additional-cni-plugins" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.445083 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.451730 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59795d48dc-ndw77"] Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.547386 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1efc9c9e-8858-4be8-bfec-386307b14e7d-client-ca\") pod \"1efc9c9e-8858-4be8-bfec-386307b14e7d\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.547450 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c52h\" (UniqueName: \"kubernetes.io/projected/1efc9c9e-8858-4be8-bfec-386307b14e7d-kube-api-access-9c52h\") pod \"1efc9c9e-8858-4be8-bfec-386307b14e7d\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.547478 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-client-ca\") pod \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.547508 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7q8l\" (UniqueName: \"kubernetes.io/projected/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-kube-api-access-f7q8l\") pod \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.547817 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-proxy-ca-bundles\") pod \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.547862 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-cni-sysctl-allowlist\") pod \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.547888 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1efc9c9e-8858-4be8-bfec-386307b14e7d-config\") pod \"1efc9c9e-8858-4be8-bfec-386307b14e7d\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.547914 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j5l7l\" (UniqueName: \"kubernetes.io/projected/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-kube-api-access-j5l7l\") pod \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.547932 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-config\") pod \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.547954 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1efc9c9e-8858-4be8-bfec-386307b14e7d-serving-cert\") pod \"1efc9c9e-8858-4be8-bfec-386307b14e7d\" (UID: \"1efc9c9e-8858-4be8-bfec-386307b14e7d\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.547975 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-serving-cert\") pod \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\" (UID: \"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.547989 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-ready\") pod \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.548017 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-tuning-conf-dir\") pod \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\" (UID: \"5206b83e-eee7-4ad4-bba2-e0c306b84aaf\") " Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.548083 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-config\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.548105 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-proxy-ca-bundles\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.548124 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lp8fn\" (UniqueName: \"kubernetes.io/projected/c4d1dda4-e584-4d86-aa39-a88007941c19-kube-api-access-lp8fn\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.548154 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-client-ca\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.548174 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4d1dda4-e584-4d86-aa39-a88007941c19-serving-cert\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.549019 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" (UID: "ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.549287 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-ready" (OuterVolumeSpecName: "ready") pod "5206b83e-eee7-4ad4-bba2-e0c306b84aaf" (UID: "5206b83e-eee7-4ad4-bba2-e0c306b84aaf"). InnerVolumeSpecName "ready". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.549316 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-tuning-conf-dir" (OuterVolumeSpecName: "tuning-conf-dir") pod "5206b83e-eee7-4ad4-bba2-e0c306b84aaf" (UID: "5206b83e-eee7-4ad4-bba2-e0c306b84aaf"). InnerVolumeSpecName "tuning-conf-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.551800 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "5206b83e-eee7-4ad4-bba2-e0c306b84aaf" (UID: "5206b83e-eee7-4ad4-bba2-e0c306b84aaf"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.552209 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1efc9c9e-8858-4be8-bfec-386307b14e7d-client-ca" (OuterVolumeSpecName: "client-ca") pod "1efc9c9e-8858-4be8-bfec-386307b14e7d" (UID: "1efc9c9e-8858-4be8-bfec-386307b14e7d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.552522 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-client-ca" (OuterVolumeSpecName: "client-ca") pod "ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" (UID: "ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.552771 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1efc9c9e-8858-4be8-bfec-386307b14e7d-config" (OuterVolumeSpecName: "config") pod "1efc9c9e-8858-4be8-bfec-386307b14e7d" (UID: "1efc9c9e-8858-4be8-bfec-386307b14e7d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.552948 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-config" (OuterVolumeSpecName: "config") pod "ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" (UID: "ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.553167 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b"] Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.553731 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" (UID: "ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.553737 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1efc9c9e-8858-4be8-bfec-386307b14e7d-kube-api-access-9c52h" (OuterVolumeSpecName: "kube-api-access-9c52h") pod "1efc9c9e-8858-4be8-bfec-386307b14e7d" (UID: "1efc9c9e-8858-4be8-bfec-386307b14e7d"). InnerVolumeSpecName "kube-api-access-9c52h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.554232 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-kube-api-access-f7q8l" (OuterVolumeSpecName: "kube-api-access-f7q8l") pod "5206b83e-eee7-4ad4-bba2-e0c306b84aaf" (UID: "5206b83e-eee7-4ad4-bba2-e0c306b84aaf"). InnerVolumeSpecName "kube-api-access-f7q8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.557626 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-kube-api-access-j5l7l" (OuterVolumeSpecName: "kube-api-access-j5l7l") pod "ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" (UID: "ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e"). InnerVolumeSpecName "kube-api-access-j5l7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.558533 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1efc9c9e-8858-4be8-bfec-386307b14e7d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1efc9c9e-8858-4be8-bfec-386307b14e7d" (UID: "1efc9c9e-8858-4be8-bfec-386307b14e7d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649113 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-client-ca\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649423 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4d1dda4-e584-4d86-aa39-a88007941c19-serving-cert\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649495 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-config\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649517 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-proxy-ca-bundles\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649537 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lp8fn\" (UniqueName: \"kubernetes.io/projected/c4d1dda4-e584-4d86-aa39-a88007941c19-kube-api-access-lp8fn\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649581 4930 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649595 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1efc9c9e-8858-4be8-bfec-386307b14e7d-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649606 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j5l7l\" (UniqueName: \"kubernetes.io/projected/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-kube-api-access-j5l7l\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649615 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649622 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1efc9c9e-8858-4be8-bfec-386307b14e7d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649633 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649640 4930 reconciler_common.go:293] "Volume detached for volume \"ready\" (UniqueName: \"kubernetes.io/empty-dir/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-ready\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649648 4930 reconciler_common.go:293] "Volume detached for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-tuning-conf-dir\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649656 4930 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1efc9c9e-8858-4be8-bfec-386307b14e7d-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649663 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c52h\" (UniqueName: \"kubernetes.io/projected/1efc9c9e-8858-4be8-bfec-386307b14e7d-kube-api-access-9c52h\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649674 4930 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649683 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7q8l\" (UniqueName: \"kubernetes.io/projected/5206b83e-eee7-4ad4-bba2-e0c306b84aaf-kube-api-access-f7q8l\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.649691 4930 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.650082 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-client-ca\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.651037 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-config\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.651179 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-proxy-ca-bundles\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.653211 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4d1dda4-e584-4d86-aa39-a88007941c19-serving-cert\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.672206 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lp8fn\" (UniqueName: \"kubernetes.io/projected/c4d1dda4-e584-4d86-aa39-a88007941c19-kube-api-access-lp8fn\") pod \"controller-manager-59795d48dc-ndw77\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.762246 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.762737 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f9946846-zp4pk" event={"ID":"ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e","Type":"ContainerDied","Data":"acd9f944ff6cf51ead8c5fe84925dc7226a9f0eaa39b95f7e39fa9ebbbcb1e49"} Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.762788 4930 scope.go:117] "RemoveContainer" containerID="2b72c49600ccf2bfd81c39ec804a8661af9d5bb30be58d6564f2ec08dad29736" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.766465 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrbn4" event={"ID":"bf6e93be-455f-45e4-a51b-d851591c31ed","Type":"ContainerStarted","Data":"3e4db90cb31b93c9af1338a9605a4d8528c3d6fd86c55011861513726d18c5cc"} Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.768730 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" event={"ID":"1efc9c9e-8858-4be8-bfec-386307b14e7d","Type":"ContainerDied","Data":"b68be83e4daedfd8d21fd8ec9730c5766519bc3000ddca45a5e78d1d44d34ff9"} Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.768763 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.770685 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_cni-sysctl-allowlist-ds-qnw8j_5206b83e-eee7-4ad4-bba2-e0c306b84aaf/kube-multus-additional-cni-plugins/0.log" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.770830 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" event={"ID":"5206b83e-eee7-4ad4-bba2-e0c306b84aaf","Type":"ContainerDied","Data":"66ae88c2acc110e16791269c7861aa1ce230b4f99d990633a957594692689665"} Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.770857 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-multus/cni-sysctl-allowlist-ds-qnw8j" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.772491 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" event={"ID":"58844174-f3d8-4618-a396-78d721d893b8","Type":"ContainerStarted","Data":"818e40fcb7b2a9b3918b664f7ccb63ffebdf47c83c92db171b851960d36fd850"} Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.772531 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" event={"ID":"58844174-f3d8-4618-a396-78d721d893b8","Type":"ContainerStarted","Data":"47393e1540d4189900c0112d5a5dddce8fdef2a2efea7635f2d665496203e89d"} Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.773687 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-tnq4n" podUID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" Mar 13 09:15:09 crc kubenswrapper[4930]: E0313 09:15:09.774115 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-zp7px" podUID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.807147 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.809716 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.886412 4930 scope.go:117] "RemoveContainer" containerID="e983976c294bc96dd5a09833a6ccdd77e91b0b72c6d0e46244729872e40ec29c" Mar 13 09:15:09 crc kubenswrapper[4930]: W0313 09:15:09.887243 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod74eb4334_fb07_42c4_b59f_7a07546af2e9.slice/crio-b4550b44c7b1d600009ff471349f43b3fa422b0d7a40323219ba4316072391a2 WatchSource:0}: Error finding container b4550b44c7b1d600009ff471349f43b3fa422b0d7a40323219ba4316072391a2: Status 404 returned error can't find the container with id b4550b44c7b1d600009ff471349f43b3fa422b0d7a40323219ba4316072391a2 Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.919867 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-qnw8j"] Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.919949 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-multus/cni-sysctl-allowlist-ds-qnw8j"] Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.933185 4930 scope.go:117] "RemoveContainer" containerID="a94666bf59c9e6762c9e7d0bc260b8b7919b23aeaf8403047b9ff29730125486" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.943655 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw"] Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.948253 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7976bf54d-zhqmw"] Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.954731 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5f9946846-zp4pk"] Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.957242 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5f9946846-zp4pk"] Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.980448 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1efc9c9e-8858-4be8-bfec-386307b14e7d" path="/var/lib/kubelet/pods/1efc9c9e-8858-4be8-bfec-386307b14e7d/volumes" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.983251 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5206b83e-eee7-4ad4-bba2-e0c306b84aaf" path="/var/lib/kubelet/pods/5206b83e-eee7-4ad4-bba2-e0c306b84aaf/volumes" Mar 13 09:15:09 crc kubenswrapper[4930]: I0313 09:15:09.984617 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e" path="/var/lib/kubelet/pods/ac6d8c7d-0e26-43fb-9508-51cddc2b5f9e/volumes" Mar 13 09:15:10 crc kubenswrapper[4930]: I0313 09:15:10.245384 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-59795d48dc-ndw77"] Mar 13 09:15:10 crc kubenswrapper[4930]: I0313 09:15:10.791424 4930 generic.go:334] "Generic (PLEG): container finished" podID="bf6e93be-455f-45e4-a51b-d851591c31ed" containerID="3e4db90cb31b93c9af1338a9605a4d8528c3d6fd86c55011861513726d18c5cc" exitCode=0 Mar 13 09:15:10 crc kubenswrapper[4930]: I0313 09:15:10.792009 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrbn4" event={"ID":"bf6e93be-455f-45e4-a51b-d851591c31ed","Type":"ContainerDied","Data":"3e4db90cb31b93c9af1338a9605a4d8528c3d6fd86c55011861513726d18c5cc"} Mar 13 09:15:10 crc kubenswrapper[4930]: I0313 09:15:10.811688 4930 generic.go:334] "Generic (PLEG): container finished" podID="58844174-f3d8-4618-a396-78d721d893b8" containerID="818e40fcb7b2a9b3918b664f7ccb63ffebdf47c83c92db171b851960d36fd850" exitCode=0 Mar 13 09:15:10 crc kubenswrapper[4930]: I0313 09:15:10.811744 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" event={"ID":"58844174-f3d8-4618-a396-78d721d893b8","Type":"ContainerDied","Data":"818e40fcb7b2a9b3918b664f7ccb63ffebdf47c83c92db171b851960d36fd850"} Mar 13 09:15:10 crc kubenswrapper[4930]: I0313 09:15:10.819386 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"74eb4334-fb07-42c4-b59f-7a07546af2e9","Type":"ContainerStarted","Data":"e37ef9bcfae63d390b6b5860faf0fe559cba98ee80213fe91573697609ae79f8"} Mar 13 09:15:10 crc kubenswrapper[4930]: I0313 09:15:10.819419 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"74eb4334-fb07-42c4-b59f-7a07546af2e9","Type":"ContainerStarted","Data":"b4550b44c7b1d600009ff471349f43b3fa422b0d7a40323219ba4316072391a2"} Mar 13 09:15:10 crc kubenswrapper[4930]: I0313 09:15:10.823392 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" event={"ID":"c4d1dda4-e584-4d86-aa39-a88007941c19","Type":"ContainerStarted","Data":"2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358"} Mar 13 09:15:10 crc kubenswrapper[4930]: I0313 09:15:10.823460 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" event={"ID":"c4d1dda4-e584-4d86-aa39-a88007941c19","Type":"ContainerStarted","Data":"8104fbe2702194b47c70b49010eeb83f56e1a7e3b8b154b19ee8f91ab17fd6aa"} Mar 13 09:15:10 crc kubenswrapper[4930]: I0313 09:15:10.824103 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:10 crc kubenswrapper[4930]: I0313 09:15:10.827874 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:10 crc kubenswrapper[4930]: I0313 09:15:10.864221 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" podStartSLOduration=19.864205138 podStartE2EDuration="19.864205138s" podCreationTimestamp="2026-03-13 09:14:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:15:10.86353333 +0000 UTC m=+151.613448007" watchObservedRunningTime="2026-03-13 09:15:10.864205138 +0000 UTC m=+151.614119815" Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.124808 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.183143 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58844174-f3d8-4618-a396-78d721d893b8-config-volume\") pod \"58844174-f3d8-4618-a396-78d721d893b8\" (UID: \"58844174-f3d8-4618-a396-78d721d893b8\") " Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.183216 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58844174-f3d8-4618-a396-78d721d893b8-secret-volume\") pod \"58844174-f3d8-4618-a396-78d721d893b8\" (UID: \"58844174-f3d8-4618-a396-78d721d893b8\") " Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.183261 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p8k95\" (UniqueName: \"kubernetes.io/projected/58844174-f3d8-4618-a396-78d721d893b8-kube-api-access-p8k95\") pod \"58844174-f3d8-4618-a396-78d721d893b8\" (UID: \"58844174-f3d8-4618-a396-78d721d893b8\") " Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.183892 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/58844174-f3d8-4618-a396-78d721d893b8-config-volume" (OuterVolumeSpecName: "config-volume") pod "58844174-f3d8-4618-a396-78d721d893b8" (UID: "58844174-f3d8-4618-a396-78d721d893b8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.188537 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58844174-f3d8-4618-a396-78d721d893b8-kube-api-access-p8k95" (OuterVolumeSpecName: "kube-api-access-p8k95") pod "58844174-f3d8-4618-a396-78d721d893b8" (UID: "58844174-f3d8-4618-a396-78d721d893b8"). InnerVolumeSpecName "kube-api-access-p8k95". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.188637 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58844174-f3d8-4618-a396-78d721d893b8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "58844174-f3d8-4618-a396-78d721d893b8" (UID: "58844174-f3d8-4618-a396-78d721d893b8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.284726 4930 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/58844174-f3d8-4618-a396-78d721d893b8-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.284769 4930 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/58844174-f3d8-4618-a396-78d721d893b8-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.284784 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p8k95\" (UniqueName: \"kubernetes.io/projected/58844174-f3d8-4618-a396-78d721d893b8-kube-api-access-p8k95\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.829307 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.829544 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b" event={"ID":"58844174-f3d8-4618-a396-78d721d893b8","Type":"ContainerDied","Data":"47393e1540d4189900c0112d5a5dddce8fdef2a2efea7635f2d665496203e89d"} Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.830030 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="47393e1540d4189900c0112d5a5dddce8fdef2a2efea7635f2d665496203e89d" Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.833797 4930 generic.go:334] "Generic (PLEG): container finished" podID="74eb4334-fb07-42c4-b59f-7a07546af2e9" containerID="e37ef9bcfae63d390b6b5860faf0fe559cba98ee80213fe91573697609ae79f8" exitCode=0 Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.833895 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"74eb4334-fb07-42c4-b59f-7a07546af2e9","Type":"ContainerDied","Data":"e37ef9bcfae63d390b6b5860faf0fe559cba98ee80213fe91573697609ae79f8"} Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.839181 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrbn4" event={"ID":"bf6e93be-455f-45e4-a51b-d851591c31ed","Type":"ContainerStarted","Data":"f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e"} Mar 13 09:15:11 crc kubenswrapper[4930]: I0313 09:15:11.866556 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-zrbn4" podStartSLOduration=2.70160652 podStartE2EDuration="39.86652635s" podCreationTimestamp="2026-03-13 09:14:32 +0000 UTC" firstStartedPulling="2026-03-13 09:14:34.0788802 +0000 UTC m=+114.828794877" lastFinishedPulling="2026-03-13 09:15:11.24380002 +0000 UTC m=+151.993714707" observedRunningTime="2026-03-13 09:15:11.864343322 +0000 UTC m=+152.614258019" watchObservedRunningTime="2026-03-13 09:15:11.86652635 +0000 UTC m=+152.616441147" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.003349 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-59795d48dc-ndw77"] Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.023369 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s"] Mar 13 09:15:12 crc kubenswrapper[4930]: E0313 09:15:12.023658 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="58844174-f3d8-4618-a396-78d721d893b8" containerName="collect-profiles" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.023672 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="58844174-f3d8-4618-a396-78d721d893b8" containerName="collect-profiles" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.023777 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="58844174-f3d8-4618-a396-78d721d893b8" containerName="collect-profiles" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.024110 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.028716 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.028855 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.028976 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.031660 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.032135 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.032254 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.050277 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s"] Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.096726 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s"] Mar 13 09:15:12 crc kubenswrapper[4930]: E0313 09:15:12.097095 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[client-ca config kube-api-access-f8mgz serving-cert], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" podUID="2ea8c5d6-0e7c-49ef-949d-e728f725261c" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.100550 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8mgz\" (UniqueName: \"kubernetes.io/projected/2ea8c5d6-0e7c-49ef-949d-e728f725261c-kube-api-access-f8mgz\") pod \"route-controller-manager-678f99bd5d-7pp5s\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.100606 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ea8c5d6-0e7c-49ef-949d-e728f725261c-config\") pod \"route-controller-manager-678f99bd5d-7pp5s\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.100679 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ea8c5d6-0e7c-49ef-949d-e728f725261c-serving-cert\") pod \"route-controller-manager-678f99bd5d-7pp5s\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.100697 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ea8c5d6-0e7c-49ef-949d-e728f725261c-client-ca\") pod \"route-controller-manager-678f99bd5d-7pp5s\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.201475 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ea8c5d6-0e7c-49ef-949d-e728f725261c-serving-cert\") pod \"route-controller-manager-678f99bd5d-7pp5s\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.201516 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ea8c5d6-0e7c-49ef-949d-e728f725261c-client-ca\") pod \"route-controller-manager-678f99bd5d-7pp5s\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.201556 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8mgz\" (UniqueName: \"kubernetes.io/projected/2ea8c5d6-0e7c-49ef-949d-e728f725261c-kube-api-access-f8mgz\") pod \"route-controller-manager-678f99bd5d-7pp5s\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.201575 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ea8c5d6-0e7c-49ef-949d-e728f725261c-config\") pod \"route-controller-manager-678f99bd5d-7pp5s\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.202532 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ea8c5d6-0e7c-49ef-949d-e728f725261c-client-ca\") pod \"route-controller-manager-678f99bd5d-7pp5s\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.202671 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ea8c5d6-0e7c-49ef-949d-e728f725261c-config\") pod \"route-controller-manager-678f99bd5d-7pp5s\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.208665 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ea8c5d6-0e7c-49ef-949d-e728f725261c-serving-cert\") pod \"route-controller-manager-678f99bd5d-7pp5s\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.209262 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.216787 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8mgz\" (UniqueName: \"kubernetes.io/projected/2ea8c5d6-0e7c-49ef-949d-e728f725261c-kube-api-access-f8mgz\") pod \"route-controller-manager-678f99bd5d-7pp5s\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.302387 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/74eb4334-fb07-42c4-b59f-7a07546af2e9-kubelet-dir\") pod \"74eb4334-fb07-42c4-b59f-7a07546af2e9\" (UID: \"74eb4334-fb07-42c4-b59f-7a07546af2e9\") " Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.302880 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74eb4334-fb07-42c4-b59f-7a07546af2e9-kube-api-access\") pod \"74eb4334-fb07-42c4-b59f-7a07546af2e9\" (UID: \"74eb4334-fb07-42c4-b59f-7a07546af2e9\") " Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.302539 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/74eb4334-fb07-42c4-b59f-7a07546af2e9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "74eb4334-fb07-42c4-b59f-7a07546af2e9" (UID: "74eb4334-fb07-42c4-b59f-7a07546af2e9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.303605 4930 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/74eb4334-fb07-42c4-b59f-7a07546af2e9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.306027 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74eb4334-fb07-42c4-b59f-7a07546af2e9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "74eb4334-fb07-42c4-b59f-7a07546af2e9" (UID: "74eb4334-fb07-42c4-b59f-7a07546af2e9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.405422 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/74eb4334-fb07-42c4-b59f-7a07546af2e9-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.846876 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"74eb4334-fb07-42c4-b59f-7a07546af2e9","Type":"ContainerDied","Data":"b4550b44c7b1d600009ff471349f43b3fa422b0d7a40323219ba4316072391a2"} Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.846947 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4550b44c7b1d600009ff471349f43b3fa422b0d7a40323219ba4316072391a2" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.846960 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.846884 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.868844 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.913396 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ea8c5d6-0e7c-49ef-949d-e728f725261c-client-ca\") pod \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.913522 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ea8c5d6-0e7c-49ef-949d-e728f725261c-config\") pod \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.913562 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ea8c5d6-0e7c-49ef-949d-e728f725261c-serving-cert\") pod \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.913617 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8mgz\" (UniqueName: \"kubernetes.io/projected/2ea8c5d6-0e7c-49ef-949d-e728f725261c-kube-api-access-f8mgz\") pod \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\" (UID: \"2ea8c5d6-0e7c-49ef-949d-e728f725261c\") " Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.915104 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ea8c5d6-0e7c-49ef-949d-e728f725261c-config" (OuterVolumeSpecName: "config") pod "2ea8c5d6-0e7c-49ef-949d-e728f725261c" (UID: "2ea8c5d6-0e7c-49ef-949d-e728f725261c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.915417 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ea8c5d6-0e7c-49ef-949d-e728f725261c-client-ca" (OuterVolumeSpecName: "client-ca") pod "2ea8c5d6-0e7c-49ef-949d-e728f725261c" (UID: "2ea8c5d6-0e7c-49ef-949d-e728f725261c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.919378 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea8c5d6-0e7c-49ef-949d-e728f725261c-kube-api-access-f8mgz" (OuterVolumeSpecName: "kube-api-access-f8mgz") pod "2ea8c5d6-0e7c-49ef-949d-e728f725261c" (UID: "2ea8c5d6-0e7c-49ef-949d-e728f725261c"). InnerVolumeSpecName "kube-api-access-f8mgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:12 crc kubenswrapper[4930]: I0313 09:15:12.919477 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2ea8c5d6-0e7c-49ef-949d-e728f725261c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "2ea8c5d6-0e7c-49ef-949d-e728f725261c" (UID: "2ea8c5d6-0e7c-49ef-949d-e728f725261c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:13 crc kubenswrapper[4930]: I0313 09:15:13.014847 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8mgz\" (UniqueName: \"kubernetes.io/projected/2ea8c5d6-0e7c-49ef-949d-e728f725261c-kube-api-access-f8mgz\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:13 crc kubenswrapper[4930]: I0313 09:15:13.014880 4930 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/2ea8c5d6-0e7c-49ef-949d-e728f725261c-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:13 crc kubenswrapper[4930]: I0313 09:15:13.014899 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2ea8c5d6-0e7c-49ef-949d-e728f725261c-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:13 crc kubenswrapper[4930]: I0313 09:15:13.014908 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2ea8c5d6-0e7c-49ef-949d-e728f725261c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:13 crc kubenswrapper[4930]: I0313 09:15:13.176296 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:15:13 crc kubenswrapper[4930]: I0313 09:15:13.176755 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:15:13 crc kubenswrapper[4930]: I0313 09:15:13.851255 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s" Mar 13 09:15:13 crc kubenswrapper[4930]: I0313 09:15:13.851396 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" podUID="c4d1dda4-e584-4d86-aa39-a88007941c19" containerName="controller-manager" containerID="cri-o://2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358" gracePeriod=30 Mar 13 09:15:13 crc kubenswrapper[4930]: I0313 09:15:13.890497 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s"] Mar 13 09:15:13 crc kubenswrapper[4930]: I0313 09:15:13.895551 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-678f99bd5d-7pp5s"] Mar 13 09:15:13 crc kubenswrapper[4930]: I0313 09:15:13.980452 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ea8c5d6-0e7c-49ef-949d-e728f725261c" path="/var/lib/kubelet/pods/2ea8c5d6-0e7c-49ef-949d-e728f725261c/volumes" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.025046 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b"] Mar 13 09:15:14 crc kubenswrapper[4930]: E0313 09:15:14.025224 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74eb4334-fb07-42c4-b59f-7a07546af2e9" containerName="pruner" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.025235 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="74eb4334-fb07-42c4-b59f-7a07546af2e9" containerName="pruner" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.025326 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="74eb4334-fb07-42c4-b59f-7a07546af2e9" containerName="pruner" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.025679 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.028692 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.028862 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.028879 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.029061 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.029174 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.031892 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.039010 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b"] Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.127753 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4db37829-ac34-445d-8a50-93c3616d5123-client-ca\") pod \"route-controller-manager-bb656f78c-7gh5b\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.127859 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4db37829-ac34-445d-8a50-93c3616d5123-serving-cert\") pod \"route-controller-manager-bb656f78c-7gh5b\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.127907 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4db37829-ac34-445d-8a50-93c3616d5123-config\") pod \"route-controller-manager-bb656f78c-7gh5b\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.128073 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stbdw\" (UniqueName: \"kubernetes.io/projected/4db37829-ac34-445d-8a50-93c3616d5123-kube-api-access-stbdw\") pod \"route-controller-manager-bb656f78c-7gh5b\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.228981 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4db37829-ac34-445d-8a50-93c3616d5123-config\") pod \"route-controller-manager-bb656f78c-7gh5b\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.229013 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4db37829-ac34-445d-8a50-93c3616d5123-client-ca\") pod \"route-controller-manager-bb656f78c-7gh5b\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.229028 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4db37829-ac34-445d-8a50-93c3616d5123-serving-cert\") pod \"route-controller-manager-bb656f78c-7gh5b\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.229075 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stbdw\" (UniqueName: \"kubernetes.io/projected/4db37829-ac34-445d-8a50-93c3616d5123-kube-api-access-stbdw\") pod \"route-controller-manager-bb656f78c-7gh5b\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.231283 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4db37829-ac34-445d-8a50-93c3616d5123-client-ca\") pod \"route-controller-manager-bb656f78c-7gh5b\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.231298 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4db37829-ac34-445d-8a50-93c3616d5123-config\") pod \"route-controller-manager-bb656f78c-7gh5b\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.249363 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stbdw\" (UniqueName: \"kubernetes.io/projected/4db37829-ac34-445d-8a50-93c3616d5123-kube-api-access-stbdw\") pod \"route-controller-manager-bb656f78c-7gh5b\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.271016 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4db37829-ac34-445d-8a50-93c3616d5123-serving-cert\") pod \"route-controller-manager-bb656f78c-7gh5b\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.304487 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-zrbn4" podUID="bf6e93be-455f-45e4-a51b-d851591c31ed" containerName="registry-server" probeResult="failure" output=< Mar 13 09:15:14 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:15:14 crc kubenswrapper[4930]: > Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.342196 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.394814 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.430948 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-proxy-ca-bundles\") pod \"c4d1dda4-e584-4d86-aa39-a88007941c19\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.431353 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4d1dda4-e584-4d86-aa39-a88007941c19-serving-cert\") pod \"c4d1dda4-e584-4d86-aa39-a88007941c19\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.431390 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lp8fn\" (UniqueName: \"kubernetes.io/projected/c4d1dda4-e584-4d86-aa39-a88007941c19-kube-api-access-lp8fn\") pod \"c4d1dda4-e584-4d86-aa39-a88007941c19\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.431465 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-client-ca\") pod \"c4d1dda4-e584-4d86-aa39-a88007941c19\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.431512 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-config\") pod \"c4d1dda4-e584-4d86-aa39-a88007941c19\" (UID: \"c4d1dda4-e584-4d86-aa39-a88007941c19\") " Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.431660 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "c4d1dda4-e584-4d86-aa39-a88007941c19" (UID: "c4d1dda4-e584-4d86-aa39-a88007941c19"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.432039 4930 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.432082 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-client-ca" (OuterVolumeSpecName: "client-ca") pod "c4d1dda4-e584-4d86-aa39-a88007941c19" (UID: "c4d1dda4-e584-4d86-aa39-a88007941c19"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.432177 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-config" (OuterVolumeSpecName: "config") pod "c4d1dda4-e584-4d86-aa39-a88007941c19" (UID: "c4d1dda4-e584-4d86-aa39-a88007941c19"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.436910 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c4d1dda4-e584-4d86-aa39-a88007941c19-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "c4d1dda4-e584-4d86-aa39-a88007941c19" (UID: "c4d1dda4-e584-4d86-aa39-a88007941c19"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.437562 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4d1dda4-e584-4d86-aa39-a88007941c19-kube-api-access-lp8fn" (OuterVolumeSpecName: "kube-api-access-lp8fn") pod "c4d1dda4-e584-4d86-aa39-a88007941c19" (UID: "c4d1dda4-e584-4d86-aa39-a88007941c19"). InnerVolumeSpecName "kube-api-access-lp8fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.511089 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b"] Mar 13 09:15:14 crc kubenswrapper[4930]: W0313 09:15:14.517018 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4db37829_ac34_445d_8a50_93c3616d5123.slice/crio-5c1ce8fc558cfcd02bf6da47822540de380699bf0b77e24b6fc3b4fa8dfe4233 WatchSource:0}: Error finding container 5c1ce8fc558cfcd02bf6da47822540de380699bf0b77e24b6fc3b4fa8dfe4233: Status 404 returned error can't find the container with id 5c1ce8fc558cfcd02bf6da47822540de380699bf0b77e24b6fc3b4fa8dfe4233 Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.533617 4930 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.533650 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4d1dda4-e584-4d86-aa39-a88007941c19-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.533662 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c4d1dda4-e584-4d86-aa39-a88007941c19-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.533675 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lp8fn\" (UniqueName: \"kubernetes.io/projected/c4d1dda4-e584-4d86-aa39-a88007941c19-kube-api-access-lp8fn\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.858450 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" event={"ID":"4db37829-ac34-445d-8a50-93c3616d5123","Type":"ContainerStarted","Data":"b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6"} Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.858727 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" event={"ID":"4db37829-ac34-445d-8a50-93c3616d5123","Type":"ContainerStarted","Data":"5c1ce8fc558cfcd02bf6da47822540de380699bf0b77e24b6fc3b4fa8dfe4233"} Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.858747 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.860321 4930 generic.go:334] "Generic (PLEG): container finished" podID="c4d1dda4-e584-4d86-aa39-a88007941c19" containerID="2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358" exitCode=0 Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.860391 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" event={"ID":"c4d1dda4-e584-4d86-aa39-a88007941c19","Type":"ContainerDied","Data":"2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358"} Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.860446 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" event={"ID":"c4d1dda4-e584-4d86-aa39-a88007941c19","Type":"ContainerDied","Data":"8104fbe2702194b47c70b49010eeb83f56e1a7e3b8b154b19ee8f91ab17fd6aa"} Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.860480 4930 scope.go:117] "RemoveContainer" containerID="2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.860513 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-59795d48dc-ndw77" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.874724 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" podStartSLOduration=2.874708178 podStartE2EDuration="2.874708178s" podCreationTimestamp="2026-03-13 09:15:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:15:14.872072357 +0000 UTC m=+155.621987054" watchObservedRunningTime="2026-03-13 09:15:14.874708178 +0000 UTC m=+155.624622845" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.878227 4930 scope.go:117] "RemoveContainer" containerID="2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358" Mar 13 09:15:14 crc kubenswrapper[4930]: E0313 09:15:14.878827 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358\": container with ID starting with 2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358 not found: ID does not exist" containerID="2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.878866 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358"} err="failed to get container status \"2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358\": rpc error: code = NotFound desc = could not find container \"2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358\": container with ID starting with 2019298787e9de2d27ac0a2687368c1afe71a18b85061dc26bcbe00fc63ab358 not found: ID does not exist" Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.898983 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-59795d48dc-ndw77"] Mar 13 09:15:14 crc kubenswrapper[4930]: I0313 09:15:14.907811 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-59795d48dc-ndw77"] Mar 13 09:15:15 crc kubenswrapper[4930]: I0313 09:15:15.086542 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:15 crc kubenswrapper[4930]: I0313 09:15:15.978056 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4d1dda4-e584-4d86-aa39-a88007941c19" path="/var/lib/kubelet/pods/c4d1dda4-e584-4d86-aa39-a88007941c19/volumes" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.022406 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k"] Mar 13 09:15:17 crc kubenswrapper[4930]: E0313 09:15:17.023640 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4d1dda4-e584-4d86-aa39-a88007941c19" containerName="controller-manager" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.023736 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4d1dda4-e584-4d86-aa39-a88007941c19" containerName="controller-manager" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.023951 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4d1dda4-e584-4d86-aa39-a88007941c19" containerName="controller-manager" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.024489 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.026215 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.027555 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.029391 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.029634 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.029758 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.031256 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.032499 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k"] Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.038021 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.062257 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-proxy-ca-bundles\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.062941 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldtpk\" (UniqueName: \"kubernetes.io/projected/093129f4-dd75-4117-8b2c-b3bba58bcf61-kube-api-access-ldtpk\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.063016 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-config\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.063119 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/093129f4-dd75-4117-8b2c-b3bba58bcf61-serving-cert\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.063197 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-client-ca\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.164881 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-proxy-ca-bundles\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.164964 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldtpk\" (UniqueName: \"kubernetes.io/projected/093129f4-dd75-4117-8b2c-b3bba58bcf61-kube-api-access-ldtpk\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.165017 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-config\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.165063 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/093129f4-dd75-4117-8b2c-b3bba58bcf61-serving-cert\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.165110 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-client-ca\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.167065 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-client-ca\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.168709 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-proxy-ca-bundles\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.170232 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-config\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.174659 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/093129f4-dd75-4117-8b2c-b3bba58bcf61-serving-cert\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.184739 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldtpk\" (UniqueName: \"kubernetes.io/projected/093129f4-dd75-4117-8b2c-b3bba58bcf61-kube-api-access-ldtpk\") pod \"controller-manager-7cd66cb78f-2gc6k\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.321148 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.344860 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.709883 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k"] Mar 13 09:15:17 crc kubenswrapper[4930]: W0313 09:15:17.718796 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod093129f4_dd75_4117_8b2c_b3bba58bcf61.slice/crio-1c32d2fd1e0c1a9cafe40c3590405ec30effb1028a6e589b95e5fb8ca8822386 WatchSource:0}: Error finding container 1c32d2fd1e0c1a9cafe40c3590405ec30effb1028a6e589b95e5fb8ca8822386: Status 404 returned error can't find the container with id 1c32d2fd1e0c1a9cafe40c3590405ec30effb1028a6e589b95e5fb8ca8822386 Mar 13 09:15:17 crc kubenswrapper[4930]: I0313 09:15:17.883676 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" event={"ID":"093129f4-dd75-4117-8b2c-b3bba58bcf61","Type":"ContainerStarted","Data":"1c32d2fd1e0c1a9cafe40c3590405ec30effb1028a6e589b95e5fb8ca8822386"} Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.623642 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.624665 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.626260 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.629271 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.634352 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.689628 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9340acf3-4897-4fb5-8977-61de13adde82-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9340acf3-4897-4fb5-8977-61de13adde82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.689683 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9340acf3-4897-4fb5-8977-61de13adde82-kube-api-access\") pod \"installer-9-crc\" (UID: \"9340acf3-4897-4fb5-8977-61de13adde82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.689829 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9340acf3-4897-4fb5-8977-61de13adde82-var-lock\") pod \"installer-9-crc\" (UID: \"9340acf3-4897-4fb5-8977-61de13adde82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.791266 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9340acf3-4897-4fb5-8977-61de13adde82-kube-api-access\") pod \"installer-9-crc\" (UID: \"9340acf3-4897-4fb5-8977-61de13adde82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.791333 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9340acf3-4897-4fb5-8977-61de13adde82-var-lock\") pod \"installer-9-crc\" (UID: \"9340acf3-4897-4fb5-8977-61de13adde82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.791366 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9340acf3-4897-4fb5-8977-61de13adde82-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9340acf3-4897-4fb5-8977-61de13adde82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.791421 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9340acf3-4897-4fb5-8977-61de13adde82-kubelet-dir\") pod \"installer-9-crc\" (UID: \"9340acf3-4897-4fb5-8977-61de13adde82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.791972 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9340acf3-4897-4fb5-8977-61de13adde82-var-lock\") pod \"installer-9-crc\" (UID: \"9340acf3-4897-4fb5-8977-61de13adde82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.814884 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9340acf3-4897-4fb5-8977-61de13adde82-kube-api-access\") pod \"installer-9-crc\" (UID: \"9340acf3-4897-4fb5-8977-61de13adde82\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.893362 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" event={"ID":"093129f4-dd75-4117-8b2c-b3bba58bcf61","Type":"ContainerStarted","Data":"32b03bedb16fc187f72dbd22c1695302d383595b5d0a5324c9b19b95fb56fe50"} Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.894913 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.902069 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.915424 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" podStartSLOduration=6.915403896 podStartE2EDuration="6.915403896s" podCreationTimestamp="2026-03-13 09:15:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:15:18.912075597 +0000 UTC m=+159.661990314" watchObservedRunningTime="2026-03-13 09:15:18.915403896 +0000 UTC m=+159.665318603" Mar 13 09:15:18 crc kubenswrapper[4930]: I0313 09:15:18.944901 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:19 crc kubenswrapper[4930]: I0313 09:15:19.132053 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 13 09:15:19 crc kubenswrapper[4930]: I0313 09:15:19.901545 4930 generic.go:334] "Generic (PLEG): container finished" podID="d36d0eb9-1490-4b68-b637-631b7a17c10a" containerID="7482fa0af538897dc3e8691fecfe600b77fcc9e0e9736433f2d754f62e33ac3c" exitCode=0 Mar 13 09:15:19 crc kubenswrapper[4930]: I0313 09:15:19.901612 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bgrn" event={"ID":"d36d0eb9-1490-4b68-b637-631b7a17c10a","Type":"ContainerDied","Data":"7482fa0af538897dc3e8691fecfe600b77fcc9e0e9736433f2d754f62e33ac3c"} Mar 13 09:15:19 crc kubenswrapper[4930]: I0313 09:15:19.903913 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9340acf3-4897-4fb5-8977-61de13adde82","Type":"ContainerStarted","Data":"fcca095ffc1041d2cb023d40521974b76498e8120d0ec351c9dd752604e485a8"} Mar 13 09:15:19 crc kubenswrapper[4930]: I0313 09:15:19.903963 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9340acf3-4897-4fb5-8977-61de13adde82","Type":"ContainerStarted","Data":"96d0ff923c0dc44258bf1a8efaaf016429dea5d490d212c4e33b631ed4ddf1c2"} Mar 13 09:15:19 crc kubenswrapper[4930]: I0313 09:15:19.934538 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.934517488 podStartE2EDuration="1.934517488s" podCreationTimestamp="2026-03-13 09:15:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:15:19.932650988 +0000 UTC m=+160.682565685" watchObservedRunningTime="2026-03-13 09:15:19.934517488 +0000 UTC m=+160.684432165" Mar 13 09:15:21 crc kubenswrapper[4930]: I0313 09:15:21.366844 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nsh7c"] Mar 13 09:15:21 crc kubenswrapper[4930]: I0313 09:15:21.916105 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78vkd" event={"ID":"980ffd8b-d762-4aac-8037-ccab15a76309","Type":"ContainerStarted","Data":"309da0572c60d5ca9cb3208df623c0d205ddbf632db7ee213372b8304898f075"} Mar 13 09:15:21 crc kubenswrapper[4930]: I0313 09:15:21.919184 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bgrn" event={"ID":"d36d0eb9-1490-4b68-b637-631b7a17c10a","Type":"ContainerStarted","Data":"9c3c3ec907cc440ab472be1203e8e8647065fc2e38f8c8aee4f7cc07f912ee1f"} Mar 13 09:15:21 crc kubenswrapper[4930]: I0313 09:15:21.931635 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvrjq" event={"ID":"5933708c-4e7f-4567-bd1e-df3dfe85fcd5","Type":"ContainerStarted","Data":"cddfdd5aa1f76cb6428df177e8b597ff13e0216ebc2def1eed47177c18b7b8e5"} Mar 13 09:15:21 crc kubenswrapper[4930]: I0313 09:15:21.973222 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-4bgrn" podStartSLOduration=1.608751143 podStartE2EDuration="48.973205232s" podCreationTimestamp="2026-03-13 09:14:33 +0000 UTC" firstStartedPulling="2026-03-13 09:14:34.055867342 +0000 UTC m=+114.805782029" lastFinishedPulling="2026-03-13 09:15:21.420321441 +0000 UTC m=+162.170236118" observedRunningTime="2026-03-13 09:15:21.972114193 +0000 UTC m=+162.722028870" watchObservedRunningTime="2026-03-13 09:15:21.973205232 +0000 UTC m=+162.723119909" Mar 13 09:15:22 crc kubenswrapper[4930]: I0313 09:15:22.940110 4930 generic.go:334] "Generic (PLEG): container finished" podID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" containerID="8f161c659679fa505c313dbc765abfbd9b4e0a1f3ec631c391e20da704d76bab" exitCode=0 Mar 13 09:15:22 crc kubenswrapper[4930]: I0313 09:15:22.940181 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zp7px" event={"ID":"27ab43e5-6d65-4d21-a496-7c724e62b9b1","Type":"ContainerDied","Data":"8f161c659679fa505c313dbc765abfbd9b4e0a1f3ec631c391e20da704d76bab"} Mar 13 09:15:22 crc kubenswrapper[4930]: I0313 09:15:22.953990 4930 generic.go:334] "Generic (PLEG): container finished" podID="980ffd8b-d762-4aac-8037-ccab15a76309" containerID="309da0572c60d5ca9cb3208df623c0d205ddbf632db7ee213372b8304898f075" exitCode=0 Mar 13 09:15:22 crc kubenswrapper[4930]: I0313 09:15:22.954111 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78vkd" event={"ID":"980ffd8b-d762-4aac-8037-ccab15a76309","Type":"ContainerDied","Data":"309da0572c60d5ca9cb3208df623c0d205ddbf632db7ee213372b8304898f075"} Mar 13 09:15:22 crc kubenswrapper[4930]: I0313 09:15:22.983290 4930 generic.go:334] "Generic (PLEG): container finished" podID="085d31a1-d613-4e2f-a992-42bfc9b66413" containerID="f5f5b1abe368a8c35d77094d68c958aba293c94236448f088cc8e8498463ae52" exitCode=0 Mar 13 09:15:22 crc kubenswrapper[4930]: I0313 09:15:22.983419 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b56p5" event={"ID":"085d31a1-d613-4e2f-a992-42bfc9b66413","Type":"ContainerDied","Data":"f5f5b1abe368a8c35d77094d68c958aba293c94236448f088cc8e8498463ae52"} Mar 13 09:15:23 crc kubenswrapper[4930]: I0313 09:15:23.028675 4930 generic.go:334] "Generic (PLEG): container finished" podID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerID="cddfdd5aa1f76cb6428df177e8b597ff13e0216ebc2def1eed47177c18b7b8e5" exitCode=0 Mar 13 09:15:23 crc kubenswrapper[4930]: I0313 09:15:23.029018 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvrjq" event={"ID":"5933708c-4e7f-4567-bd1e-df3dfe85fcd5","Type":"ContainerDied","Data":"cddfdd5aa1f76cb6428df177e8b597ff13e0216ebc2def1eed47177c18b7b8e5"} Mar 13 09:15:23 crc kubenswrapper[4930]: I0313 09:15:23.220873 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:15:23 crc kubenswrapper[4930]: I0313 09:15:23.262719 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:15:23 crc kubenswrapper[4930]: E0313 09:15:23.535785 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod342671c3_b1f8_4c61_a0da_abe7dcdfc367.slice/crio-conmon-f0976d49697dba856874e4cf8f556c8b0961bc51aa9cb18c1ecda66745359dfe.scope\": RecentStats: unable to find data in memory cache]" Mar 13 09:15:23 crc kubenswrapper[4930]: I0313 09:15:23.658373 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:15:23 crc kubenswrapper[4930]: I0313 09:15:23.659338 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:15:23 crc kubenswrapper[4930]: I0313 09:15:23.701769 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:15:24 crc kubenswrapper[4930]: I0313 09:15:24.035739 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zp7px" event={"ID":"27ab43e5-6d65-4d21-a496-7c724e62b9b1","Type":"ContainerStarted","Data":"e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d"} Mar 13 09:15:24 crc kubenswrapper[4930]: I0313 09:15:24.037245 4930 generic.go:334] "Generic (PLEG): container finished" podID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" containerID="f0976d49697dba856874e4cf8f556c8b0961bc51aa9cb18c1ecda66745359dfe" exitCode=0 Mar 13 09:15:24 crc kubenswrapper[4930]: I0313 09:15:24.037296 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tnq4n" event={"ID":"342671c3-b1f8-4c61-a0da-abe7dcdfc367","Type":"ContainerDied","Data":"f0976d49697dba856874e4cf8f556c8b0961bc51aa9cb18c1ecda66745359dfe"} Mar 13 09:15:24 crc kubenswrapper[4930]: I0313 09:15:24.039228 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78vkd" event={"ID":"980ffd8b-d762-4aac-8037-ccab15a76309","Type":"ContainerStarted","Data":"cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982"} Mar 13 09:15:24 crc kubenswrapper[4930]: I0313 09:15:24.041323 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b56p5" event={"ID":"085d31a1-d613-4e2f-a992-42bfc9b66413","Type":"ContainerStarted","Data":"f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd"} Mar 13 09:15:24 crc kubenswrapper[4930]: I0313 09:15:24.043960 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvrjq" event={"ID":"5933708c-4e7f-4567-bd1e-df3dfe85fcd5","Type":"ContainerStarted","Data":"79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2"} Mar 13 09:15:24 crc kubenswrapper[4930]: I0313 09:15:24.054844 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zp7px" podStartSLOduration=1.6980118229999999 podStartE2EDuration="51.054827036s" podCreationTimestamp="2026-03-13 09:14:33 +0000 UTC" firstStartedPulling="2026-03-13 09:14:34.067584127 +0000 UTC m=+114.817498804" lastFinishedPulling="2026-03-13 09:15:23.42439934 +0000 UTC m=+164.174314017" observedRunningTime="2026-03-13 09:15:24.052414661 +0000 UTC m=+164.802329338" watchObservedRunningTime="2026-03-13 09:15:24.054827036 +0000 UTC m=+164.804741713" Mar 13 09:15:24 crc kubenswrapper[4930]: I0313 09:15:24.067581 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cvrjq" podStartSLOduration=2.853519985 podStartE2EDuration="49.067567927s" podCreationTimestamp="2026-03-13 09:14:35 +0000 UTC" firstStartedPulling="2026-03-13 09:14:37.369356055 +0000 UTC m=+118.119270732" lastFinishedPulling="2026-03-13 09:15:23.583403997 +0000 UTC m=+164.333318674" observedRunningTime="2026-03-13 09:15:24.06730918 +0000 UTC m=+164.817223857" watchObservedRunningTime="2026-03-13 09:15:24.067567927 +0000 UTC m=+164.817482604" Mar 13 09:15:24 crc kubenswrapper[4930]: I0313 09:15:24.088062 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-78vkd" podStartSLOduration=2.964099686 podStartE2EDuration="48.088047855s" podCreationTimestamp="2026-03-13 09:14:36 +0000 UTC" firstStartedPulling="2026-03-13 09:14:38.426688011 +0000 UTC m=+119.176602688" lastFinishedPulling="2026-03-13 09:15:23.55063618 +0000 UTC m=+164.300550857" observedRunningTime="2026-03-13 09:15:24.085161198 +0000 UTC m=+164.835075875" watchObservedRunningTime="2026-03-13 09:15:24.088047855 +0000 UTC m=+164.837962532" Mar 13 09:15:24 crc kubenswrapper[4930]: I0313 09:15:24.101873 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b56p5" podStartSLOduration=2.798170644 podStartE2EDuration="50.101858045s" podCreationTimestamp="2026-03-13 09:14:34 +0000 UTC" firstStartedPulling="2026-03-13 09:14:36.185798502 +0000 UTC m=+116.935713179" lastFinishedPulling="2026-03-13 09:15:23.489485903 +0000 UTC m=+164.239400580" observedRunningTime="2026-03-13 09:15:24.100844798 +0000 UTC m=+164.850759485" watchObservedRunningTime="2026-03-13 09:15:24.101858045 +0000 UTC m=+164.851772722" Mar 13 09:15:25 crc kubenswrapper[4930]: I0313 09:15:25.051452 4930 generic.go:334] "Generic (PLEG): container finished" podID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" containerID="0cee50f464d2bfe4e54f8ca95a56a307af9deea6dfe4f881f483a31257f05eec" exitCode=0 Mar 13 09:15:25 crc kubenswrapper[4930]: I0313 09:15:25.051585 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87r56" event={"ID":"6deff2ec-a8cb-4a0a-b752-bbba80f3512d","Type":"ContainerDied","Data":"0cee50f464d2bfe4e54f8ca95a56a307af9deea6dfe4f881f483a31257f05eec"} Mar 13 09:15:25 crc kubenswrapper[4930]: I0313 09:15:25.053868 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tnq4n" event={"ID":"342671c3-b1f8-4c61-a0da-abe7dcdfc367","Type":"ContainerStarted","Data":"dcf841d84dca2fde34daf6d3dd2b33890ac1c4ae49782c2d6977ba7a83f68c5b"} Mar 13 09:15:25 crc kubenswrapper[4930]: I0313 09:15:25.095273 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tnq4n" podStartSLOduration=2.671869268 podStartE2EDuration="53.095252548s" podCreationTimestamp="2026-03-13 09:14:32 +0000 UTC" firstStartedPulling="2026-03-13 09:14:34.049879547 +0000 UTC m=+114.799794224" lastFinishedPulling="2026-03-13 09:15:24.473262827 +0000 UTC m=+165.223177504" observedRunningTime="2026-03-13 09:15:25.092293529 +0000 UTC m=+165.842208206" watchObservedRunningTime="2026-03-13 09:15:25.095252548 +0000 UTC m=+165.845167235" Mar 13 09:15:25 crc kubenswrapper[4930]: I0313 09:15:25.183202 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:15:25 crc kubenswrapper[4930]: I0313 09:15:25.183262 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:15:26 crc kubenswrapper[4930]: I0313 09:15:26.060504 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87r56" event={"ID":"6deff2ec-a8cb-4a0a-b752-bbba80f3512d","Type":"ContainerStarted","Data":"2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31"} Mar 13 09:15:26 crc kubenswrapper[4930]: I0313 09:15:26.083909 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-87r56" podStartSLOduration=2.685387532 podStartE2EDuration="51.083891854s" podCreationTimestamp="2026-03-13 09:14:35 +0000 UTC" firstStartedPulling="2026-03-13 09:14:37.384636431 +0000 UTC m=+118.134551098" lastFinishedPulling="2026-03-13 09:15:25.783140743 +0000 UTC m=+166.533055420" observedRunningTime="2026-03-13 09:15:26.083692348 +0000 UTC m=+166.833607025" watchObservedRunningTime="2026-03-13 09:15:26.083891854 +0000 UTC m=+166.833806551" Mar 13 09:15:26 crc kubenswrapper[4930]: I0313 09:15:26.180213 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:15:26 crc kubenswrapper[4930]: I0313 09:15:26.180264 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:15:26 crc kubenswrapper[4930]: I0313 09:15:26.228691 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-b56p5" podUID="085d31a1-d613-4e2f-a992-42bfc9b66413" containerName="registry-server" probeResult="failure" output=< Mar 13 09:15:26 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:15:26 crc kubenswrapper[4930]: > Mar 13 09:15:26 crc kubenswrapper[4930]: I0313 09:15:26.641167 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:15:26 crc kubenswrapper[4930]: I0313 09:15:26.641221 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:15:27 crc kubenswrapper[4930]: I0313 09:15:27.213807 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cvrjq" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerName="registry-server" probeResult="failure" output=< Mar 13 09:15:27 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:15:27 crc kubenswrapper[4930]: > Mar 13 09:15:27 crc kubenswrapper[4930]: I0313 09:15:27.708408 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-78vkd" podUID="980ffd8b-d762-4aac-8037-ccab15a76309" containerName="registry-server" probeResult="failure" output=< Mar 13 09:15:27 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:15:27 crc kubenswrapper[4930]: > Mar 13 09:15:31 crc kubenswrapper[4930]: I0313 09:15:31.978970 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k"] Mar 13 09:15:31 crc kubenswrapper[4930]: I0313 09:15:31.979496 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" podUID="093129f4-dd75-4117-8b2c-b3bba58bcf61" containerName="controller-manager" containerID="cri-o://32b03bedb16fc187f72dbd22c1695302d383595b5d0a5324c9b19b95fb56fe50" gracePeriod=30 Mar 13 09:15:31 crc kubenswrapper[4930]: I0313 09:15:31.985288 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b"] Mar 13 09:15:31 crc kubenswrapper[4930]: I0313 09:15:31.985524 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" podUID="4db37829-ac34-445d-8a50-93c3616d5123" containerName="route-controller-manager" containerID="cri-o://b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6" gracePeriod=30 Mar 13 09:15:32 crc kubenswrapper[4930]: I0313 09:15:32.999256 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:32.999780 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.045471 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.056334 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.059418 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4db37829-ac34-445d-8a50-93c3616d5123-client-ca\") pod \"4db37829-ac34-445d-8a50-93c3616d5123\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.059507 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4db37829-ac34-445d-8a50-93c3616d5123-serving-cert\") pod \"4db37829-ac34-445d-8a50-93c3616d5123\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.059554 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4db37829-ac34-445d-8a50-93c3616d5123-config\") pod \"4db37829-ac34-445d-8a50-93c3616d5123\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.059637 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stbdw\" (UniqueName: \"kubernetes.io/projected/4db37829-ac34-445d-8a50-93c3616d5123-kube-api-access-stbdw\") pod \"4db37829-ac34-445d-8a50-93c3616d5123\" (UID: \"4db37829-ac34-445d-8a50-93c3616d5123\") " Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.060318 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4db37829-ac34-445d-8a50-93c3616d5123-client-ca" (OuterVolumeSpecName: "client-ca") pod "4db37829-ac34-445d-8a50-93c3616d5123" (UID: "4db37829-ac34-445d-8a50-93c3616d5123"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.060343 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4db37829-ac34-445d-8a50-93c3616d5123-config" (OuterVolumeSpecName: "config") pod "4db37829-ac34-445d-8a50-93c3616d5123" (UID: "4db37829-ac34-445d-8a50-93c3616d5123"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.065864 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4db37829-ac34-445d-8a50-93c3616d5123-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4db37829-ac34-445d-8a50-93c3616d5123" (UID: "4db37829-ac34-445d-8a50-93c3616d5123"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.065864 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4db37829-ac34-445d-8a50-93c3616d5123-kube-api-access-stbdw" (OuterVolumeSpecName: "kube-api-access-stbdw") pod "4db37829-ac34-445d-8a50-93c3616d5123" (UID: "4db37829-ac34-445d-8a50-93c3616d5123"). InnerVolumeSpecName "kube-api-access-stbdw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.111600 4930 generic.go:334] "Generic (PLEG): container finished" podID="093129f4-dd75-4117-8b2c-b3bba58bcf61" containerID="32b03bedb16fc187f72dbd22c1695302d383595b5d0a5324c9b19b95fb56fe50" exitCode=0 Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.112041 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" event={"ID":"093129f4-dd75-4117-8b2c-b3bba58bcf61","Type":"ContainerDied","Data":"32b03bedb16fc187f72dbd22c1695302d383595b5d0a5324c9b19b95fb56fe50"} Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.113939 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7"] Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.114501 4930 generic.go:334] "Generic (PLEG): container finished" podID="4db37829-ac34-445d-8a50-93c3616d5123" containerID="b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6" exitCode=0 Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.114665 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" Mar 13 09:15:33 crc kubenswrapper[4930]: E0313 09:15:33.115878 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4db37829-ac34-445d-8a50-93c3616d5123" containerName="route-controller-manager" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.115910 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="4db37829-ac34-445d-8a50-93c3616d5123" containerName="route-controller-manager" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.116171 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="4db37829-ac34-445d-8a50-93c3616d5123" containerName="route-controller-manager" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.116717 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" event={"ID":"4db37829-ac34-445d-8a50-93c3616d5123","Type":"ContainerDied","Data":"b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6"} Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.116758 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b" event={"ID":"4db37829-ac34-445d-8a50-93c3616d5123","Type":"ContainerDied","Data":"5c1ce8fc558cfcd02bf6da47822540de380699bf0b77e24b6fc3b4fa8dfe4233"} Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.116797 4930 scope.go:117] "RemoveContainer" containerID="b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.116998 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.120021 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.120782 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.120941 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.121628 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.121670 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.121781 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.131729 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7"] Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.160884 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-config\") pod \"route-controller-manager-65986ccb59-kqxc7\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.161059 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-serving-cert\") pod \"route-controller-manager-65986ccb59-kqxc7\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.161102 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbrq9\" (UniqueName: \"kubernetes.io/projected/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-kube-api-access-zbrq9\") pod \"route-controller-manager-65986ccb59-kqxc7\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.161151 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-client-ca\") pod \"route-controller-manager-65986ccb59-kqxc7\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.161213 4930 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4db37829-ac34-445d-8a50-93c3616d5123-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.161235 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4db37829-ac34-445d-8a50-93c3616d5123-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.161254 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4db37829-ac34-445d-8a50-93c3616d5123-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.161272 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stbdw\" (UniqueName: \"kubernetes.io/projected/4db37829-ac34-445d-8a50-93c3616d5123-kube-api-access-stbdw\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.166656 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.168140 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b"] Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.168265 4930 scope.go:117] "RemoveContainer" containerID="b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6" Mar 13 09:15:33 crc kubenswrapper[4930]: E0313 09:15:33.168583 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6\": container with ID starting with b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6 not found: ID does not exist" containerID="b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.168622 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6"} err="failed to get container status \"b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6\": rpc error: code = NotFound desc = could not find container \"b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6\": container with ID starting with b45d276c8e159ab14239553e2f46106602d06ecf8357e7d25cc56fe64d5e36d6 not found: ID does not exist" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.183001 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bb656f78c-7gh5b"] Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.189884 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.261551 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ldtpk\" (UniqueName: \"kubernetes.io/projected/093129f4-dd75-4117-8b2c-b3bba58bcf61-kube-api-access-ldtpk\") pod \"093129f4-dd75-4117-8b2c-b3bba58bcf61\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.261617 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/093129f4-dd75-4117-8b2c-b3bba58bcf61-serving-cert\") pod \"093129f4-dd75-4117-8b2c-b3bba58bcf61\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.261647 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-config\") pod \"093129f4-dd75-4117-8b2c-b3bba58bcf61\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.261674 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-client-ca\") pod \"093129f4-dd75-4117-8b2c-b3bba58bcf61\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.261692 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-proxy-ca-bundles\") pod \"093129f4-dd75-4117-8b2c-b3bba58bcf61\" (UID: \"093129f4-dd75-4117-8b2c-b3bba58bcf61\") " Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.261777 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-serving-cert\") pod \"route-controller-manager-65986ccb59-kqxc7\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.261801 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbrq9\" (UniqueName: \"kubernetes.io/projected/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-kube-api-access-zbrq9\") pod \"route-controller-manager-65986ccb59-kqxc7\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.261823 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-client-ca\") pod \"route-controller-manager-65986ccb59-kqxc7\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.261851 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-config\") pod \"route-controller-manager-65986ccb59-kqxc7\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.262532 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "093129f4-dd75-4117-8b2c-b3bba58bcf61" (UID: "093129f4-dd75-4117-8b2c-b3bba58bcf61"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.262592 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-config" (OuterVolumeSpecName: "config") pod "093129f4-dd75-4117-8b2c-b3bba58bcf61" (UID: "093129f4-dd75-4117-8b2c-b3bba58bcf61"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.262586 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-client-ca" (OuterVolumeSpecName: "client-ca") pod "093129f4-dd75-4117-8b2c-b3bba58bcf61" (UID: "093129f4-dd75-4117-8b2c-b3bba58bcf61"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.263279 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-config\") pod \"route-controller-manager-65986ccb59-kqxc7\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.264272 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/093129f4-dd75-4117-8b2c-b3bba58bcf61-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "093129f4-dd75-4117-8b2c-b3bba58bcf61" (UID: "093129f4-dd75-4117-8b2c-b3bba58bcf61"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.264757 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/093129f4-dd75-4117-8b2c-b3bba58bcf61-kube-api-access-ldtpk" (OuterVolumeSpecName: "kube-api-access-ldtpk") pod "093129f4-dd75-4117-8b2c-b3bba58bcf61" (UID: "093129f4-dd75-4117-8b2c-b3bba58bcf61"). InnerVolumeSpecName "kube-api-access-ldtpk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.265267 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-serving-cert\") pod \"route-controller-manager-65986ccb59-kqxc7\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.267071 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-client-ca\") pod \"route-controller-manager-65986ccb59-kqxc7\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.278090 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbrq9\" (UniqueName: \"kubernetes.io/projected/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-kube-api-access-zbrq9\") pod \"route-controller-manager-65986ccb59-kqxc7\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.362775 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.362812 4930 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.362825 4930 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/093129f4-dd75-4117-8b2c-b3bba58bcf61-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.362838 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ldtpk\" (UniqueName: \"kubernetes.io/projected/093129f4-dd75-4117-8b2c-b3bba58bcf61-kube-api-access-ldtpk\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.362848 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/093129f4-dd75-4117-8b2c-b3bba58bcf61-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.436542 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.436877 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.469287 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.484275 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.682907 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7"] Mar 13 09:15:33 crc kubenswrapper[4930]: W0313 09:15:33.699061 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7beccb5c_af60_4ea3_9519_c3f6c09c61b2.slice/crio-b0fe084d3838339ab3dc3f0a98409984c0ef704ae2f6479f390d3a260ff4c4e6 WatchSource:0}: Error finding container b0fe084d3838339ab3dc3f0a98409984c0ef704ae2f6479f390d3a260ff4c4e6: Status 404 returned error can't find the container with id b0fe084d3838339ab3dc3f0a98409984c0ef704ae2f6479f390d3a260ff4c4e6 Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.700013 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:15:33 crc kubenswrapper[4930]: I0313 09:15:33.980790 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4db37829-ac34-445d-8a50-93c3616d5123" path="/var/lib/kubelet/pods/4db37829-ac34-445d-8a50-93c3616d5123/volumes" Mar 13 09:15:34 crc kubenswrapper[4930]: I0313 09:15:34.122167 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" event={"ID":"7beccb5c-af60-4ea3-9519-c3f6c09c61b2","Type":"ContainerStarted","Data":"603eeceb959d84ae1c602a15aa35b4534ed9ff63b5fced93a329ff0db85e1c12"} Mar 13 09:15:34 crc kubenswrapper[4930]: I0313 09:15:34.122226 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" event={"ID":"7beccb5c-af60-4ea3-9519-c3f6c09c61b2","Type":"ContainerStarted","Data":"b0fe084d3838339ab3dc3f0a98409984c0ef704ae2f6479f390d3a260ff4c4e6"} Mar 13 09:15:34 crc kubenswrapper[4930]: I0313 09:15:34.122261 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:34 crc kubenswrapper[4930]: I0313 09:15:34.123535 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" event={"ID":"093129f4-dd75-4117-8b2c-b3bba58bcf61","Type":"ContainerDied","Data":"1c32d2fd1e0c1a9cafe40c3590405ec30effb1028a6e589b95e5fb8ca8822386"} Mar 13 09:15:34 crc kubenswrapper[4930]: I0313 09:15:34.123582 4930 scope.go:117] "RemoveContainer" containerID="32b03bedb16fc187f72dbd22c1695302d383595b5d0a5324c9b19b95fb56fe50" Mar 13 09:15:34 crc kubenswrapper[4930]: I0313 09:15:34.123691 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k" Mar 13 09:15:34 crc kubenswrapper[4930]: I0313 09:15:34.159418 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" podStartSLOduration=2.159395392 podStartE2EDuration="2.159395392s" podCreationTimestamp="2026-03-13 09:15:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:15:34.14550373 +0000 UTC m=+174.895418407" watchObservedRunningTime="2026-03-13 09:15:34.159395392 +0000 UTC m=+174.909310059" Mar 13 09:15:34 crc kubenswrapper[4930]: I0313 09:15:34.162879 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k"] Mar 13 09:15:34 crc kubenswrapper[4930]: I0313 09:15:34.163800 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7cd66cb78f-2gc6k"] Mar 13 09:15:34 crc kubenswrapper[4930]: I0313 09:15:34.172097 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:15:34 crc kubenswrapper[4930]: I0313 09:15:34.189579 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:35 crc kubenswrapper[4930]: I0313 09:15:35.079722 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zp7px"] Mar 13 09:15:35 crc kubenswrapper[4930]: I0313 09:15:35.218031 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:15:35 crc kubenswrapper[4930]: I0313 09:15:35.253207 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:15:35 crc kubenswrapper[4930]: I0313 09:15:35.610308 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:15:35 crc kubenswrapper[4930]: I0313 09:15:35.610382 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:15:35 crc kubenswrapper[4930]: I0313 09:15:35.663983 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:15:35 crc kubenswrapper[4930]: I0313 09:15:35.683077 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4bgrn"] Mar 13 09:15:35 crc kubenswrapper[4930]: I0313 09:15:35.683322 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-4bgrn" podUID="d36d0eb9-1490-4b68-b637-631b7a17c10a" containerName="registry-server" containerID="cri-o://9c3c3ec907cc440ab472be1203e8e8647065fc2e38f8c8aee4f7cc07f912ee1f" gracePeriod=2 Mar 13 09:15:35 crc kubenswrapper[4930]: I0313 09:15:35.982798 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="093129f4-dd75-4117-8b2c-b3bba58bcf61" path="/var/lib/kubelet/pods/093129f4-dd75-4117-8b2c-b3bba58bcf61/volumes" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.049864 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6bf94df96d-jcbsn"] Mar 13 09:15:36 crc kubenswrapper[4930]: E0313 09:15:36.050195 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="093129f4-dd75-4117-8b2c-b3bba58bcf61" containerName="controller-manager" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.050227 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="093129f4-dd75-4117-8b2c-b3bba58bcf61" containerName="controller-manager" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.050550 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="093129f4-dd75-4117-8b2c-b3bba58bcf61" containerName="controller-manager" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.051221 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.053703 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.054121 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.054581 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.055198 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.055381 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.059131 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.062836 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bf94df96d-jcbsn"] Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.066980 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.144343 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zp7px" podUID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" containerName="registry-server" containerID="cri-o://e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d" gracePeriod=2 Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.189403 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.198138 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-proxy-ca-bundles\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.198596 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncq6h\" (UniqueName: \"kubernetes.io/projected/4d3f3d99-a52a-404a-b665-5d05736e2cb7-kube-api-access-ncq6h\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.198639 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-client-ca\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.198673 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-config\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.198769 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d3f3d99-a52a-404a-b665-5d05736e2cb7-serving-cert\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.231527 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.275394 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.300682 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncq6h\" (UniqueName: \"kubernetes.io/projected/4d3f3d99-a52a-404a-b665-5d05736e2cb7-kube-api-access-ncq6h\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.300777 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-client-ca\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.300857 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-config\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.301018 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d3f3d99-a52a-404a-b665-5d05736e2cb7-serving-cert\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.301255 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-proxy-ca-bundles\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.302093 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-client-ca\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.303789 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-proxy-ca-bundles\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.305017 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-config\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.322303 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d3f3d99-a52a-404a-b665-5d05736e2cb7-serving-cert\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.336892 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncq6h\" (UniqueName: \"kubernetes.io/projected/4d3f3d99-a52a-404a-b665-5d05736e2cb7-kube-api-access-ncq6h\") pod \"controller-manager-6bf94df96d-jcbsn\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.373404 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.587221 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bf94df96d-jcbsn"] Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.681407 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:15:36 crc kubenswrapper[4930]: I0313 09:15:36.731183 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:15:37 crc kubenswrapper[4930]: I0313 09:15:37.173869 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" event={"ID":"4d3f3d99-a52a-404a-b665-5d05736e2cb7","Type":"ContainerStarted","Data":"6a73ede4ad661249e95be813375fff74fb2e10c2d7df9fd17dbc12ed95b4068a"} Mar 13 09:15:37 crc kubenswrapper[4930]: I0313 09:15:37.177151 4930 generic.go:334] "Generic (PLEG): container finished" podID="d36d0eb9-1490-4b68-b637-631b7a17c10a" containerID="9c3c3ec907cc440ab472be1203e8e8647065fc2e38f8c8aee4f7cc07f912ee1f" exitCode=0 Mar 13 09:15:37 crc kubenswrapper[4930]: I0313 09:15:37.178059 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bgrn" event={"ID":"d36d0eb9-1490-4b68-b637-631b7a17c10a","Type":"ContainerDied","Data":"9c3c3ec907cc440ab472be1203e8e8647065fc2e38f8c8aee4f7cc07f912ee1f"} Mar 13 09:15:37 crc kubenswrapper[4930]: I0313 09:15:37.747295 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:15:37 crc kubenswrapper[4930]: I0313 09:15:37.925234 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d36d0eb9-1490-4b68-b637-631b7a17c10a-catalog-content\") pod \"d36d0eb9-1490-4b68-b637-631b7a17c10a\" (UID: \"d36d0eb9-1490-4b68-b637-631b7a17c10a\") " Mar 13 09:15:37 crc kubenswrapper[4930]: I0313 09:15:37.925334 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5jlk\" (UniqueName: \"kubernetes.io/projected/d36d0eb9-1490-4b68-b637-631b7a17c10a-kube-api-access-g5jlk\") pod \"d36d0eb9-1490-4b68-b637-631b7a17c10a\" (UID: \"d36d0eb9-1490-4b68-b637-631b7a17c10a\") " Mar 13 09:15:37 crc kubenswrapper[4930]: I0313 09:15:37.925484 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d36d0eb9-1490-4b68-b637-631b7a17c10a-utilities\") pod \"d36d0eb9-1490-4b68-b637-631b7a17c10a\" (UID: \"d36d0eb9-1490-4b68-b637-631b7a17c10a\") " Mar 13 09:15:37 crc kubenswrapper[4930]: I0313 09:15:37.926863 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d36d0eb9-1490-4b68-b637-631b7a17c10a-utilities" (OuterVolumeSpecName: "utilities") pod "d36d0eb9-1490-4b68-b637-631b7a17c10a" (UID: "d36d0eb9-1490-4b68-b637-631b7a17c10a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:15:37 crc kubenswrapper[4930]: I0313 09:15:37.941746 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d36d0eb9-1490-4b68-b637-631b7a17c10a-kube-api-access-g5jlk" (OuterVolumeSpecName: "kube-api-access-g5jlk") pod "d36d0eb9-1490-4b68-b637-631b7a17c10a" (UID: "d36d0eb9-1490-4b68-b637-631b7a17c10a"). InnerVolumeSpecName "kube-api-access-g5jlk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:37 crc kubenswrapper[4930]: I0313 09:15:37.990637 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d36d0eb9-1490-4b68-b637-631b7a17c10a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d36d0eb9-1490-4b68-b637-631b7a17c10a" (UID: "d36d0eb9-1490-4b68-b637-631b7a17c10a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.026702 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5jlk\" (UniqueName: \"kubernetes.io/projected/d36d0eb9-1490-4b68-b637-631b7a17c10a-kube-api-access-g5jlk\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.026730 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d36d0eb9-1490-4b68-b637-631b7a17c10a-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.026741 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d36d0eb9-1490-4b68-b637-631b7a17c10a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.039083 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.078831 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-87r56"] Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.127360 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74xgk\" (UniqueName: \"kubernetes.io/projected/27ab43e5-6d65-4d21-a496-7c724e62b9b1-kube-api-access-74xgk\") pod \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\" (UID: \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\") " Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.127517 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27ab43e5-6d65-4d21-a496-7c724e62b9b1-utilities\") pod \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\" (UID: \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\") " Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.127560 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27ab43e5-6d65-4d21-a496-7c724e62b9b1-catalog-content\") pod \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\" (UID: \"27ab43e5-6d65-4d21-a496-7c724e62b9b1\") " Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.128336 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27ab43e5-6d65-4d21-a496-7c724e62b9b1-utilities" (OuterVolumeSpecName: "utilities") pod "27ab43e5-6d65-4d21-a496-7c724e62b9b1" (UID: "27ab43e5-6d65-4d21-a496-7c724e62b9b1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.130934 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27ab43e5-6d65-4d21-a496-7c724e62b9b1-kube-api-access-74xgk" (OuterVolumeSpecName: "kube-api-access-74xgk") pod "27ab43e5-6d65-4d21-a496-7c724e62b9b1" (UID: "27ab43e5-6d65-4d21-a496-7c724e62b9b1"). InnerVolumeSpecName "kube-api-access-74xgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.184587 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" event={"ID":"4d3f3d99-a52a-404a-b665-5d05736e2cb7","Type":"ContainerStarted","Data":"12991f6427df4be9b552f70169a9c788da303373eaa7ab78fbcf2b8a0135a15f"} Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.184803 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.187212 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-4bgrn" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.187211 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-4bgrn" event={"ID":"d36d0eb9-1490-4b68-b637-631b7a17c10a","Type":"ContainerDied","Data":"b27528e5ed93952f5748e182214de895bbdc624a57129e3b89d8d5f1a37ef85c"} Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.187279 4930 scope.go:117] "RemoveContainer" containerID="9c3c3ec907cc440ab472be1203e8e8647065fc2e38f8c8aee4f7cc07f912ee1f" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.189250 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.191565 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zp7px" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.191714 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zp7px" event={"ID":"27ab43e5-6d65-4d21-a496-7c724e62b9b1","Type":"ContainerDied","Data":"e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d"} Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.191629 4930 generic.go:334] "Generic (PLEG): container finished" podID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" containerID="e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d" exitCode=0 Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.191877 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zp7px" event={"ID":"27ab43e5-6d65-4d21-a496-7c724e62b9b1","Type":"ContainerDied","Data":"e3d5713531526a39c7ab61b70b4dedace9d92762cb903037cda74999b6b5e858"} Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.192156 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-87r56" podUID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" containerName="registry-server" containerID="cri-o://2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31" gracePeriod=2 Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.207305 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" podStartSLOduration=6.207288522 podStartE2EDuration="6.207288522s" podCreationTimestamp="2026-03-13 09:15:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:15:38.20498334 +0000 UTC m=+178.954898017" watchObservedRunningTime="2026-03-13 09:15:38.207288522 +0000 UTC m=+178.957203199" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.207676 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27ab43e5-6d65-4d21-a496-7c724e62b9b1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "27ab43e5-6d65-4d21-a496-7c724e62b9b1" (UID: "27ab43e5-6d65-4d21-a496-7c724e62b9b1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.211274 4930 scope.go:117] "RemoveContainer" containerID="7482fa0af538897dc3e8691fecfe600b77fcc9e0e9736433f2d754f62e33ac3c" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.231260 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/27ab43e5-6d65-4d21-a496-7c724e62b9b1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.231291 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74xgk\" (UniqueName: \"kubernetes.io/projected/27ab43e5-6d65-4d21-a496-7c724e62b9b1-kube-api-access-74xgk\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.231301 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/27ab43e5-6d65-4d21-a496-7c724e62b9b1-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.245246 4930 scope.go:117] "RemoveContainer" containerID="71918f546ba6672b3ec75855d8cefefc042c2bec4aae16a03fa2e37760fc83f3" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.248644 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-4bgrn"] Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.251185 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-4bgrn"] Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.263650 4930 scope.go:117] "RemoveContainer" containerID="e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.280747 4930 scope.go:117] "RemoveContainer" containerID="8f161c659679fa505c313dbc765abfbd9b4e0a1f3ec631c391e20da704d76bab" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.358046 4930 scope.go:117] "RemoveContainer" containerID="b0a0bac87521fa54656c88f068ede02f28d9a95a863f594ec3a60a541c7b47ec" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.379338 4930 scope.go:117] "RemoveContainer" containerID="e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d" Mar 13 09:15:38 crc kubenswrapper[4930]: E0313 09:15:38.379733 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d\": container with ID starting with e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d not found: ID does not exist" containerID="e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.379761 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d"} err="failed to get container status \"e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d\": rpc error: code = NotFound desc = could not find container \"e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d\": container with ID starting with e3cc0ddf9dd54a09aa3a06c902d208bdf43bc8202517a1ea922c8a3771e5cc2d not found: ID does not exist" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.379784 4930 scope.go:117] "RemoveContainer" containerID="8f161c659679fa505c313dbc765abfbd9b4e0a1f3ec631c391e20da704d76bab" Mar 13 09:15:38 crc kubenswrapper[4930]: E0313 09:15:38.379996 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f161c659679fa505c313dbc765abfbd9b4e0a1f3ec631c391e20da704d76bab\": container with ID starting with 8f161c659679fa505c313dbc765abfbd9b4e0a1f3ec631c391e20da704d76bab not found: ID does not exist" containerID="8f161c659679fa505c313dbc765abfbd9b4e0a1f3ec631c391e20da704d76bab" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.380030 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f161c659679fa505c313dbc765abfbd9b4e0a1f3ec631c391e20da704d76bab"} err="failed to get container status \"8f161c659679fa505c313dbc765abfbd9b4e0a1f3ec631c391e20da704d76bab\": rpc error: code = NotFound desc = could not find container \"8f161c659679fa505c313dbc765abfbd9b4e0a1f3ec631c391e20da704d76bab\": container with ID starting with 8f161c659679fa505c313dbc765abfbd9b4e0a1f3ec631c391e20da704d76bab not found: ID does not exist" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.380047 4930 scope.go:117] "RemoveContainer" containerID="b0a0bac87521fa54656c88f068ede02f28d9a95a863f594ec3a60a541c7b47ec" Mar 13 09:15:38 crc kubenswrapper[4930]: E0313 09:15:38.380316 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0a0bac87521fa54656c88f068ede02f28d9a95a863f594ec3a60a541c7b47ec\": container with ID starting with b0a0bac87521fa54656c88f068ede02f28d9a95a863f594ec3a60a541c7b47ec not found: ID does not exist" containerID="b0a0bac87521fa54656c88f068ede02f28d9a95a863f594ec3a60a541c7b47ec" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.380337 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0a0bac87521fa54656c88f068ede02f28d9a95a863f594ec3a60a541c7b47ec"} err="failed to get container status \"b0a0bac87521fa54656c88f068ede02f28d9a95a863f594ec3a60a541c7b47ec\": rpc error: code = NotFound desc = could not find container \"b0a0bac87521fa54656c88f068ede02f28d9a95a863f594ec3a60a541c7b47ec\": container with ID starting with b0a0bac87521fa54656c88f068ede02f28d9a95a863f594ec3a60a541c7b47ec not found: ID does not exist" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.515757 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.517205 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zp7px"] Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.521584 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zp7px"] Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.636491 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-catalog-content\") pod \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\" (UID: \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\") " Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.636652 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frkrf\" (UniqueName: \"kubernetes.io/projected/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-kube-api-access-frkrf\") pod \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\" (UID: \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\") " Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.636682 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-utilities\") pod \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\" (UID: \"6deff2ec-a8cb-4a0a-b752-bbba80f3512d\") " Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.637549 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-utilities" (OuterVolumeSpecName: "utilities") pod "6deff2ec-a8cb-4a0a-b752-bbba80f3512d" (UID: "6deff2ec-a8cb-4a0a-b752-bbba80f3512d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.640605 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-kube-api-access-frkrf" (OuterVolumeSpecName: "kube-api-access-frkrf") pod "6deff2ec-a8cb-4a0a-b752-bbba80f3512d" (UID: "6deff2ec-a8cb-4a0a-b752-bbba80f3512d"). InnerVolumeSpecName "kube-api-access-frkrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.675097 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6deff2ec-a8cb-4a0a-b752-bbba80f3512d" (UID: "6deff2ec-a8cb-4a0a-b752-bbba80f3512d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.738573 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.738627 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frkrf\" (UniqueName: \"kubernetes.io/projected/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-kube-api-access-frkrf\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:38 crc kubenswrapper[4930]: I0313 09:15:38.738651 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6deff2ec-a8cb-4a0a-b752-bbba80f3512d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.203630 4930 generic.go:334] "Generic (PLEG): container finished" podID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" containerID="2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31" exitCode=0 Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.203686 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-87r56" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.203703 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87r56" event={"ID":"6deff2ec-a8cb-4a0a-b752-bbba80f3512d","Type":"ContainerDied","Data":"2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31"} Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.204197 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-87r56" event={"ID":"6deff2ec-a8cb-4a0a-b752-bbba80f3512d","Type":"ContainerDied","Data":"a2cce1e46af2875bab02ef1a980177811e6af1077a8e9c4530f54f8c21689c88"} Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.204243 4930 scope.go:117] "RemoveContainer" containerID="2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.232796 4930 scope.go:117] "RemoveContainer" containerID="0cee50f464d2bfe4e54f8ca95a56a307af9deea6dfe4f881f483a31257f05eec" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.253466 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-87r56"] Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.253536 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-87r56"] Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.271866 4930 scope.go:117] "RemoveContainer" containerID="8e122a18c69e1590b7ac0e9eee2230a480d7b017cc44973dd5f44feac9390c3a" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.290997 4930 scope.go:117] "RemoveContainer" containerID="2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31" Mar 13 09:15:39 crc kubenswrapper[4930]: E0313 09:15:39.291547 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31\": container with ID starting with 2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31 not found: ID does not exist" containerID="2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.291580 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31"} err="failed to get container status \"2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31\": rpc error: code = NotFound desc = could not find container \"2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31\": container with ID starting with 2482a423817b1f2b8c9041fc0396d763dff8770809eb601cc4f9f7dd6cd36a31 not found: ID does not exist" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.291610 4930 scope.go:117] "RemoveContainer" containerID="0cee50f464d2bfe4e54f8ca95a56a307af9deea6dfe4f881f483a31257f05eec" Mar 13 09:15:39 crc kubenswrapper[4930]: E0313 09:15:39.292144 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0cee50f464d2bfe4e54f8ca95a56a307af9deea6dfe4f881f483a31257f05eec\": container with ID starting with 0cee50f464d2bfe4e54f8ca95a56a307af9deea6dfe4f881f483a31257f05eec not found: ID does not exist" containerID="0cee50f464d2bfe4e54f8ca95a56a307af9deea6dfe4f881f483a31257f05eec" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.292178 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0cee50f464d2bfe4e54f8ca95a56a307af9deea6dfe4f881f483a31257f05eec"} err="failed to get container status \"0cee50f464d2bfe4e54f8ca95a56a307af9deea6dfe4f881f483a31257f05eec\": rpc error: code = NotFound desc = could not find container \"0cee50f464d2bfe4e54f8ca95a56a307af9deea6dfe4f881f483a31257f05eec\": container with ID starting with 0cee50f464d2bfe4e54f8ca95a56a307af9deea6dfe4f881f483a31257f05eec not found: ID does not exist" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.292202 4930 scope.go:117] "RemoveContainer" containerID="8e122a18c69e1590b7ac0e9eee2230a480d7b017cc44973dd5f44feac9390c3a" Mar 13 09:15:39 crc kubenswrapper[4930]: E0313 09:15:39.292688 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8e122a18c69e1590b7ac0e9eee2230a480d7b017cc44973dd5f44feac9390c3a\": container with ID starting with 8e122a18c69e1590b7ac0e9eee2230a480d7b017cc44973dd5f44feac9390c3a not found: ID does not exist" containerID="8e122a18c69e1590b7ac0e9eee2230a480d7b017cc44973dd5f44feac9390c3a" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.292793 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8e122a18c69e1590b7ac0e9eee2230a480d7b017cc44973dd5f44feac9390c3a"} err="failed to get container status \"8e122a18c69e1590b7ac0e9eee2230a480d7b017cc44973dd5f44feac9390c3a\": rpc error: code = NotFound desc = could not find container \"8e122a18c69e1590b7ac0e9eee2230a480d7b017cc44973dd5f44feac9390c3a\": container with ID starting with 8e122a18c69e1590b7ac0e9eee2230a480d7b017cc44973dd5f44feac9390c3a not found: ID does not exist" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.976877 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" path="/var/lib/kubelet/pods/27ab43e5-6d65-4d21-a496-7c724e62b9b1/volumes" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.977534 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" path="/var/lib/kubelet/pods/6deff2ec-a8cb-4a0a-b752-bbba80f3512d/volumes" Mar 13 09:15:39 crc kubenswrapper[4930]: I0313 09:15:39.978202 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d36d0eb9-1490-4b68-b637-631b7a17c10a" path="/var/lib/kubelet/pods/d36d0eb9-1490-4b68-b637-631b7a17c10a/volumes" Mar 13 09:15:40 crc kubenswrapper[4930]: I0313 09:15:40.481821 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-78vkd"] Mar 13 09:15:40 crc kubenswrapper[4930]: I0313 09:15:40.482050 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-78vkd" podUID="980ffd8b-d762-4aac-8037-ccab15a76309" containerName="registry-server" containerID="cri-o://cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982" gracePeriod=2 Mar 13 09:15:40 crc kubenswrapper[4930]: I0313 09:15:40.903775 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.064576 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/980ffd8b-d762-4aac-8037-ccab15a76309-utilities\") pod \"980ffd8b-d762-4aac-8037-ccab15a76309\" (UID: \"980ffd8b-d762-4aac-8037-ccab15a76309\") " Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.064664 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rvxdj\" (UniqueName: \"kubernetes.io/projected/980ffd8b-d762-4aac-8037-ccab15a76309-kube-api-access-rvxdj\") pod \"980ffd8b-d762-4aac-8037-ccab15a76309\" (UID: \"980ffd8b-d762-4aac-8037-ccab15a76309\") " Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.064730 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/980ffd8b-d762-4aac-8037-ccab15a76309-catalog-content\") pod \"980ffd8b-d762-4aac-8037-ccab15a76309\" (UID: \"980ffd8b-d762-4aac-8037-ccab15a76309\") " Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.065905 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/980ffd8b-d762-4aac-8037-ccab15a76309-utilities" (OuterVolumeSpecName: "utilities") pod "980ffd8b-d762-4aac-8037-ccab15a76309" (UID: "980ffd8b-d762-4aac-8037-ccab15a76309"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.082681 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/980ffd8b-d762-4aac-8037-ccab15a76309-kube-api-access-rvxdj" (OuterVolumeSpecName: "kube-api-access-rvxdj") pod "980ffd8b-d762-4aac-8037-ccab15a76309" (UID: "980ffd8b-d762-4aac-8037-ccab15a76309"). InnerVolumeSpecName "kube-api-access-rvxdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.166738 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/980ffd8b-d762-4aac-8037-ccab15a76309-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.166780 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rvxdj\" (UniqueName: \"kubernetes.io/projected/980ffd8b-d762-4aac-8037-ccab15a76309-kube-api-access-rvxdj\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.216590 4930 generic.go:334] "Generic (PLEG): container finished" podID="980ffd8b-d762-4aac-8037-ccab15a76309" containerID="cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982" exitCode=0 Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.216643 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78vkd" event={"ID":"980ffd8b-d762-4aac-8037-ccab15a76309","Type":"ContainerDied","Data":"cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982"} Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.216724 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-78vkd" event={"ID":"980ffd8b-d762-4aac-8037-ccab15a76309","Type":"ContainerDied","Data":"536da4b970ca3ac794fb8d2236f2ab9e9197fa93e1ce5e4e876024bdbe57c2fa"} Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.216760 4930 scope.go:117] "RemoveContainer" containerID="cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.216669 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-78vkd" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.234688 4930 scope.go:117] "RemoveContainer" containerID="309da0572c60d5ca9cb3208df623c0d205ddbf632db7ee213372b8304898f075" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.253500 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/980ffd8b-d762-4aac-8037-ccab15a76309-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "980ffd8b-d762-4aac-8037-ccab15a76309" (UID: "980ffd8b-d762-4aac-8037-ccab15a76309"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.264637 4930 scope.go:117] "RemoveContainer" containerID="84c1e2558aa40b4e4217c5eab421123a53239ae0ed0be2363278282d1521b81e" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.268180 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/980ffd8b-d762-4aac-8037-ccab15a76309-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.283258 4930 scope.go:117] "RemoveContainer" containerID="cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982" Mar 13 09:15:41 crc kubenswrapper[4930]: E0313 09:15:41.283759 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982\": container with ID starting with cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982 not found: ID does not exist" containerID="cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.283791 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982"} err="failed to get container status \"cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982\": rpc error: code = NotFound desc = could not find container \"cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982\": container with ID starting with cfcc6d384fe438e4f52d615c8dd6e77ff3a53a13d3ba2423ae2e3785015c9982 not found: ID does not exist" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.283811 4930 scope.go:117] "RemoveContainer" containerID="309da0572c60d5ca9cb3208df623c0d205ddbf632db7ee213372b8304898f075" Mar 13 09:15:41 crc kubenswrapper[4930]: E0313 09:15:41.284073 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"309da0572c60d5ca9cb3208df623c0d205ddbf632db7ee213372b8304898f075\": container with ID starting with 309da0572c60d5ca9cb3208df623c0d205ddbf632db7ee213372b8304898f075 not found: ID does not exist" containerID="309da0572c60d5ca9cb3208df623c0d205ddbf632db7ee213372b8304898f075" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.284095 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"309da0572c60d5ca9cb3208df623c0d205ddbf632db7ee213372b8304898f075"} err="failed to get container status \"309da0572c60d5ca9cb3208df623c0d205ddbf632db7ee213372b8304898f075\": rpc error: code = NotFound desc = could not find container \"309da0572c60d5ca9cb3208df623c0d205ddbf632db7ee213372b8304898f075\": container with ID starting with 309da0572c60d5ca9cb3208df623c0d205ddbf632db7ee213372b8304898f075 not found: ID does not exist" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.284107 4930 scope.go:117] "RemoveContainer" containerID="84c1e2558aa40b4e4217c5eab421123a53239ae0ed0be2363278282d1521b81e" Mar 13 09:15:41 crc kubenswrapper[4930]: E0313 09:15:41.284506 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84c1e2558aa40b4e4217c5eab421123a53239ae0ed0be2363278282d1521b81e\": container with ID starting with 84c1e2558aa40b4e4217c5eab421123a53239ae0ed0be2363278282d1521b81e not found: ID does not exist" containerID="84c1e2558aa40b4e4217c5eab421123a53239ae0ed0be2363278282d1521b81e" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.284563 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84c1e2558aa40b4e4217c5eab421123a53239ae0ed0be2363278282d1521b81e"} err="failed to get container status \"84c1e2558aa40b4e4217c5eab421123a53239ae0ed0be2363278282d1521b81e\": rpc error: code = NotFound desc = could not find container \"84c1e2558aa40b4e4217c5eab421123a53239ae0ed0be2363278282d1521b81e\": container with ID starting with 84c1e2558aa40b4e4217c5eab421123a53239ae0ed0be2363278282d1521b81e not found: ID does not exist" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.563350 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-78vkd"] Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.567207 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-78vkd"] Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.656560 4930 ???:1] "http: TLS handshake error from 192.168.126.11:35940: no serving certificate available for the kubelet" Mar 13 09:15:41 crc kubenswrapper[4930]: I0313 09:15:41.979284 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="980ffd8b-d762-4aac-8037-ccab15a76309" path="/var/lib/kubelet/pods/980ffd8b-d762-4aac-8037-ccab15a76309/volumes" Mar 13 09:15:46 crc kubenswrapper[4930]: I0313 09:15:46.403945 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" podUID="2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" containerName="oauth-openshift" containerID="cri-o://d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124" gracePeriod=15 Mar 13 09:15:46 crc kubenswrapper[4930]: I0313 09:15:46.926010 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.043914 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htl4k\" (UniqueName: \"kubernetes.io/projected/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-kube-api-access-htl4k\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044352 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-ocp-branding-template\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044417 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-session\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044481 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-idp-0-file-data\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044519 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-router-certs\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044549 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-error\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044576 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-provider-selection\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044620 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-serving-cert\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044646 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-trusted-ca-bundle\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044683 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-service-ca\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044711 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-audit-dir\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044827 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-audit-policies\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044871 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-cliconfig\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.044899 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-login\") pod \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\" (UID: \"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53\") " Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.045395 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.045499 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.045887 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.045954 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.047730 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.053885 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.062777 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-kube-api-access-htl4k" (OuterVolumeSpecName: "kube-api-access-htl4k") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "kube-api-access-htl4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.066527 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.068069 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.068566 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.068646 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.068761 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.068978 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.069288 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" (UID: "2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147010 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147054 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147070 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147084 4930 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147097 4930 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147111 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147125 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147138 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htl4k\" (UniqueName: \"kubernetes.io/projected/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-kube-api-access-htl4k\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147151 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147163 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147175 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147189 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147201 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.147214 4930 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.249506 4930 generic.go:334] "Generic (PLEG): container finished" podID="2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" containerID="d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124" exitCode=0 Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.249549 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" event={"ID":"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53","Type":"ContainerDied","Data":"d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124"} Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.249558 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.249574 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-nsh7c" event={"ID":"2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53","Type":"ContainerDied","Data":"c311a63d469cc11d3965ebd3108f7335a282396d98fa9ec5436ffef560e9690f"} Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.249589 4930 scope.go:117] "RemoveContainer" containerID="d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.271625 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nsh7c"] Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.272699 4930 scope.go:117] "RemoveContainer" containerID="d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124" Mar 13 09:15:47 crc kubenswrapper[4930]: E0313 09:15:47.273316 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124\": container with ID starting with d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124 not found: ID does not exist" containerID="d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.273344 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124"} err="failed to get container status \"d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124\": rpc error: code = NotFound desc = could not find container \"d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124\": container with ID starting with d927c10c0c79f1a08666f0fb6770f9ab2e1226934151508b48b040aaf73b3124 not found: ID does not exist" Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.275475 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-nsh7c"] Mar 13 09:15:47 crc kubenswrapper[4930]: I0313 09:15:47.982978 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" path="/var/lib/kubelet/pods/2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53/volumes" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.064888 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-69c75b6f68-wdg64"] Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065223 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" containerName="extract-utilities" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065253 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" containerName="extract-utilities" Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065272 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" containerName="registry-server" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065286 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" containerName="registry-server" Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065306 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" containerName="extract-utilities" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065321 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" containerName="extract-utilities" Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065342 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d36d0eb9-1490-4b68-b637-631b7a17c10a" containerName="registry-server" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065355 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d36d0eb9-1490-4b68-b637-631b7a17c10a" containerName="registry-server" Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065378 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" containerName="extract-content" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065391 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" containerName="extract-content" Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065409 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" containerName="oauth-openshift" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065422 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" containerName="oauth-openshift" Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065477 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d36d0eb9-1490-4b68-b637-631b7a17c10a" containerName="extract-content" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065496 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d36d0eb9-1490-4b68-b637-631b7a17c10a" containerName="extract-content" Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065521 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" containerName="extract-content" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065537 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" containerName="extract-content" Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065559 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="980ffd8b-d762-4aac-8037-ccab15a76309" containerName="registry-server" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065576 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="980ffd8b-d762-4aac-8037-ccab15a76309" containerName="registry-server" Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065601 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="980ffd8b-d762-4aac-8037-ccab15a76309" containerName="extract-utilities" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065615 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="980ffd8b-d762-4aac-8037-ccab15a76309" containerName="extract-utilities" Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065635 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="980ffd8b-d762-4aac-8037-ccab15a76309" containerName="extract-content" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065649 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="980ffd8b-d762-4aac-8037-ccab15a76309" containerName="extract-content" Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065671 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d36d0eb9-1490-4b68-b637-631b7a17c10a" containerName="extract-utilities" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065684 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d36d0eb9-1490-4b68-b637-631b7a17c10a" containerName="extract-utilities" Mar 13 09:15:48 crc kubenswrapper[4930]: E0313 09:15:48.065704 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" containerName="registry-server" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065716 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" containerName="registry-server" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065893 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d35e51a-2f8d-4da0-9b2e-0dcf2b29cd53" containerName="oauth-openshift" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065912 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="27ab43e5-6d65-4d21-a496-7c724e62b9b1" containerName="registry-server" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065927 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6deff2ec-a8cb-4a0a-b752-bbba80f3512d" containerName="registry-server" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065951 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="980ffd8b-d762-4aac-8037-ccab15a76309" containerName="registry-server" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.065973 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d36d0eb9-1490-4b68-b637-631b7a17c10a" containerName="registry-server" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.066658 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.071557 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.072068 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.074165 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.074639 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.076819 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.077132 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.077394 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.077732 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.077957 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.078221 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.078423 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.078801 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.093640 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.096383 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.158582 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.158697 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xwr4x\" (UniqueName: \"kubernetes.io/projected/37ebb02d-bc69-46d5-b9ac-3f676826de28-kube-api-access-xwr4x\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.158745 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-serving-cert\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.158793 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.158879 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-user-template-login\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.158923 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/37ebb02d-bc69-46d5-b9ac-3f676826de28-audit-dir\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.158967 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-session\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.159016 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.159048 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-router-certs\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.159091 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.159133 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/37ebb02d-bc69-46d5-b9ac-3f676826de28-audit-policies\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.159184 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-user-template-error\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.159220 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-service-ca\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.159254 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-cliconfig\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260036 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/37ebb02d-bc69-46d5-b9ac-3f676826de28-audit-policies\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260096 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-user-template-error\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260125 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-service-ca\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260152 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-cliconfig\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260183 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260207 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xwr4x\" (UniqueName: \"kubernetes.io/projected/37ebb02d-bc69-46d5-b9ac-3f676826de28-kube-api-access-xwr4x\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260234 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-serving-cert\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260262 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260295 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-user-template-login\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260322 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/37ebb02d-bc69-46d5-b9ac-3f676826de28-audit-dir\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260350 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-session\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260380 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260405 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-router-certs\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.260449 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.264779 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.265470 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/37ebb02d-bc69-46d5-b9ac-3f676826de28-audit-policies\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.268728 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-user-template-error\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.269212 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-service-ca\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.269817 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-cliconfig\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.272353 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.276900 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-serving-cert\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.281828 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-user-template-login\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.281974 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/37ebb02d-bc69-46d5-b9ac-3f676826de28-audit-dir\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.285628 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-session\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.287101 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.290658 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-router-certs\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.304903 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xwr4x\" (UniqueName: \"kubernetes.io/projected/37ebb02d-bc69-46d5-b9ac-3f676826de28-kube-api-access-xwr4x\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.916916 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.927340 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/37ebb02d-bc69-46d5-b9ac-3f676826de28-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-69c75b6f68-wdg64\" (UID: \"37ebb02d-bc69-46d5-b9ac-3f676826de28\") " pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.927430 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-69c75b6f68-wdg64"] Mar 13 09:15:48 crc kubenswrapper[4930]: I0313 09:15:48.994819 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:49 crc kubenswrapper[4930]: I0313 09:15:49.478037 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-69c75b6f68-wdg64"] Mar 13 09:15:50 crc kubenswrapper[4930]: I0313 09:15:50.267172 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" event={"ID":"37ebb02d-bc69-46d5-b9ac-3f676826de28","Type":"ContainerStarted","Data":"34f191dfc2ed28b8f28030a7f3b536e07a821336bad5a53f8da29c19b66b5be9"} Mar 13 09:15:50 crc kubenswrapper[4930]: I0313 09:15:50.267532 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" event={"ID":"37ebb02d-bc69-46d5-b9ac-3f676826de28","Type":"ContainerStarted","Data":"8b7cfbdc8b224c3d43dec98904c3f4725bdcc950dcad6f86f0e8e7ce306b3694"} Mar 13 09:15:50 crc kubenswrapper[4930]: I0313 09:15:50.267637 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:50 crc kubenswrapper[4930]: I0313 09:15:50.273990 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 09:15:50 crc kubenswrapper[4930]: I0313 09:15:50.295515 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" podStartSLOduration=29.295495118 podStartE2EDuration="29.295495118s" podCreationTimestamp="2026-03-13 09:15:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:15:50.294708847 +0000 UTC m=+191.044623524" watchObservedRunningTime="2026-03-13 09:15:50.295495118 +0000 UTC m=+191.045409795" Mar 13 09:15:51 crc kubenswrapper[4930]: I0313 09:15:51.966426 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6bf94df96d-jcbsn"] Mar 13 09:15:51 crc kubenswrapper[4930]: I0313 09:15:51.967055 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" podUID="4d3f3d99-a52a-404a-b665-5d05736e2cb7" containerName="controller-manager" containerID="cri-o://12991f6427df4be9b552f70169a9c788da303373eaa7ab78fbcf2b8a0135a15f" gracePeriod=30 Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.066730 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7"] Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.066975 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" podUID="7beccb5c-af60-4ea3-9519-c3f6c09c61b2" containerName="route-controller-manager" containerID="cri-o://603eeceb959d84ae1c602a15aa35b4534ed9ff63b5fced93a329ff0db85e1c12" gracePeriod=30 Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.279886 4930 generic.go:334] "Generic (PLEG): container finished" podID="7beccb5c-af60-4ea3-9519-c3f6c09c61b2" containerID="603eeceb959d84ae1c602a15aa35b4534ed9ff63b5fced93a329ff0db85e1c12" exitCode=0 Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.279979 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" event={"ID":"7beccb5c-af60-4ea3-9519-c3f6c09c61b2","Type":"ContainerDied","Data":"603eeceb959d84ae1c602a15aa35b4534ed9ff63b5fced93a329ff0db85e1c12"} Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.284995 4930 generic.go:334] "Generic (PLEG): container finished" podID="4d3f3d99-a52a-404a-b665-5d05736e2cb7" containerID="12991f6427df4be9b552f70169a9c788da303373eaa7ab78fbcf2b8a0135a15f" exitCode=0 Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.285054 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" event={"ID":"4d3f3d99-a52a-404a-b665-5d05736e2cb7","Type":"ContainerDied","Data":"12991f6427df4be9b552f70169a9c788da303373eaa7ab78fbcf2b8a0135a15f"} Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.524017 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.528094 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.637341 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zbrq9\" (UniqueName: \"kubernetes.io/projected/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-kube-api-access-zbrq9\") pod \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.637392 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncq6h\" (UniqueName: \"kubernetes.io/projected/4d3f3d99-a52a-404a-b665-5d05736e2cb7-kube-api-access-ncq6h\") pod \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.637423 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-config\") pod \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.637494 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-client-ca\") pod \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.637535 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-serving-cert\") pod \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.637569 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-config\") pod \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\" (UID: \"7beccb5c-af60-4ea3-9519-c3f6c09c61b2\") " Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.637605 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-client-ca\") pod \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.637634 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d3f3d99-a52a-404a-b665-5d05736e2cb7-serving-cert\") pod \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.637668 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-proxy-ca-bundles\") pod \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\" (UID: \"4d3f3d99-a52a-404a-b665-5d05736e2cb7\") " Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.638291 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-client-ca" (OuterVolumeSpecName: "client-ca") pod "7beccb5c-af60-4ea3-9519-c3f6c09c61b2" (UID: "7beccb5c-af60-4ea3-9519-c3f6c09c61b2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.638611 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "4d3f3d99-a52a-404a-b665-5d05736e2cb7" (UID: "4d3f3d99-a52a-404a-b665-5d05736e2cb7"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.638670 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-config" (OuterVolumeSpecName: "config") pod "4d3f3d99-a52a-404a-b665-5d05736e2cb7" (UID: "4d3f3d99-a52a-404a-b665-5d05736e2cb7"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.638690 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-client-ca" (OuterVolumeSpecName: "client-ca") pod "4d3f3d99-a52a-404a-b665-5d05736e2cb7" (UID: "4d3f3d99-a52a-404a-b665-5d05736e2cb7"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.638747 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-config" (OuterVolumeSpecName: "config") pod "7beccb5c-af60-4ea3-9519-c3f6c09c61b2" (UID: "7beccb5c-af60-4ea3-9519-c3f6c09c61b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.642122 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4d3f3d99-a52a-404a-b665-5d05736e2cb7-kube-api-access-ncq6h" (OuterVolumeSpecName: "kube-api-access-ncq6h") pod "4d3f3d99-a52a-404a-b665-5d05736e2cb7" (UID: "4d3f3d99-a52a-404a-b665-5d05736e2cb7"). InnerVolumeSpecName "kube-api-access-ncq6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.643020 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-kube-api-access-zbrq9" (OuterVolumeSpecName: "kube-api-access-zbrq9") pod "7beccb5c-af60-4ea3-9519-c3f6c09c61b2" (UID: "7beccb5c-af60-4ea3-9519-c3f6c09c61b2"). InnerVolumeSpecName "kube-api-access-zbrq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.643142 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4d3f3d99-a52a-404a-b665-5d05736e2cb7-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "4d3f3d99-a52a-404a-b665-5d05736e2cb7" (UID: "4d3f3d99-a52a-404a-b665-5d05736e2cb7"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.643584 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7beccb5c-af60-4ea3-9519-c3f6c09c61b2" (UID: "7beccb5c-af60-4ea3-9519-c3f6c09c61b2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.738806 4930 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.738861 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.738874 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.738888 4930 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-client-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.738899 4930 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4d3f3d99-a52a-404a-b665-5d05736e2cb7-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.738910 4930 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.738924 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zbrq9\" (UniqueName: \"kubernetes.io/projected/7beccb5c-af60-4ea3-9519-c3f6c09c61b2-kube-api-access-zbrq9\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.738936 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncq6h\" (UniqueName: \"kubernetes.io/projected/4d3f3d99-a52a-404a-b665-5d05736e2cb7-kube-api-access-ncq6h\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:52 crc kubenswrapper[4930]: I0313 09:15:52.738951 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4d3f3d99-a52a-404a-b665-5d05736e2cb7-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.058667 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7984f94f96-clbzb"] Mar 13 09:15:53 crc kubenswrapper[4930]: E0313 09:15:53.058845 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4d3f3d99-a52a-404a-b665-5d05736e2cb7" containerName="controller-manager" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.058857 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="4d3f3d99-a52a-404a-b665-5d05736e2cb7" containerName="controller-manager" Mar 13 09:15:53 crc kubenswrapper[4930]: E0313 09:15:53.058873 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7beccb5c-af60-4ea3-9519-c3f6c09c61b2" containerName="route-controller-manager" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.058879 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="7beccb5c-af60-4ea3-9519-c3f6c09c61b2" containerName="route-controller-manager" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.058952 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="4d3f3d99-a52a-404a-b665-5d05736e2cb7" containerName="controller-manager" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.058963 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="7beccb5c-af60-4ea3-9519-c3f6c09c61b2" containerName="route-controller-manager" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.059277 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.071111 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7984f94f96-clbzb"] Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.144082 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-proxy-ca-bundles\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.144171 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-config\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.144199 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vx88z\" (UniqueName: \"kubernetes.io/projected/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-kube-api-access-vx88z\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.144237 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-client-ca\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.144344 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-serving-cert\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.245256 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-serving-cert\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.245351 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-proxy-ca-bundles\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.245390 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-config\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.245412 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vx88z\" (UniqueName: \"kubernetes.io/projected/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-kube-api-access-vx88z\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.245465 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-client-ca\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.246581 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-client-ca\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.246931 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-proxy-ca-bundles\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.247115 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-config\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.250299 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-serving-cert\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.260295 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vx88z\" (UniqueName: \"kubernetes.io/projected/73d513ce-8d01-4a5e-ad2c-a3f86c80668a-kube-api-access-vx88z\") pod \"controller-manager-7984f94f96-clbzb\" (UID: \"73d513ce-8d01-4a5e-ad2c-a3f86c80668a\") " pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.290501 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" event={"ID":"4d3f3d99-a52a-404a-b665-5d05736e2cb7","Type":"ContainerDied","Data":"6a73ede4ad661249e95be813375fff74fb2e10c2d7df9fd17dbc12ed95b4068a"} Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.290547 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bf94df96d-jcbsn" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.290582 4930 scope.go:117] "RemoveContainer" containerID="12991f6427df4be9b552f70169a9c788da303373eaa7ab78fbcf2b8a0135a15f" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.292087 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" event={"ID":"7beccb5c-af60-4ea3-9519-c3f6c09c61b2","Type":"ContainerDied","Data":"b0fe084d3838339ab3dc3f0a98409984c0ef704ae2f6479f390d3a260ff4c4e6"} Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.292174 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.305805 4930 scope.go:117] "RemoveContainer" containerID="603eeceb959d84ae1c602a15aa35b4534ed9ff63b5fced93a329ff0db85e1c12" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.321016 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7"] Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.324215 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-65986ccb59-kqxc7"] Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.333737 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6bf94df96d-jcbsn"] Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.336213 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6bf94df96d-jcbsn"] Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.375844 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.784500 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7984f94f96-clbzb"] Mar 13 09:15:53 crc kubenswrapper[4930]: W0313 09:15:53.793612 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73d513ce_8d01_4a5e_ad2c_a3f86c80668a.slice/crio-02a01d27feb77bfa412efb4fea128eb7936f5be38a00b8d38940b38f0a2337bc WatchSource:0}: Error finding container 02a01d27feb77bfa412efb4fea128eb7936f5be38a00b8d38940b38f0a2337bc: Status 404 returned error can't find the container with id 02a01d27feb77bfa412efb4fea128eb7936f5be38a00b8d38940b38f0a2337bc Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.980405 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4d3f3d99-a52a-404a-b665-5d05736e2cb7" path="/var/lib/kubelet/pods/4d3f3d99-a52a-404a-b665-5d05736e2cb7/volumes" Mar 13 09:15:53 crc kubenswrapper[4930]: I0313 09:15:53.981769 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7beccb5c-af60-4ea3-9519-c3f6c09c61b2" path="/var/lib/kubelet/pods/7beccb5c-af60-4ea3-9519-c3f6c09c61b2/volumes" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.063020 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9"] Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.063987 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.067789 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.068014 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.068173 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.068316 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.069053 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.070505 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.092554 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9"] Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.157637 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23510207-60fc-496b-b907-224360e7e887-config\") pod \"route-controller-manager-55d94fcf7b-v9cp9\" (UID: \"23510207-60fc-496b-b907-224360e7e887\") " pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.157797 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23510207-60fc-496b-b907-224360e7e887-client-ca\") pod \"route-controller-manager-55d94fcf7b-v9cp9\" (UID: \"23510207-60fc-496b-b907-224360e7e887\") " pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.158126 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23510207-60fc-496b-b907-224360e7e887-serving-cert\") pod \"route-controller-manager-55d94fcf7b-v9cp9\" (UID: \"23510207-60fc-496b-b907-224360e7e887\") " pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.158184 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm5r4\" (UniqueName: \"kubernetes.io/projected/23510207-60fc-496b-b907-224360e7e887-kube-api-access-xm5r4\") pod \"route-controller-manager-55d94fcf7b-v9cp9\" (UID: \"23510207-60fc-496b-b907-224360e7e887\") " pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.259537 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23510207-60fc-496b-b907-224360e7e887-config\") pod \"route-controller-manager-55d94fcf7b-v9cp9\" (UID: \"23510207-60fc-496b-b907-224360e7e887\") " pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.259603 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23510207-60fc-496b-b907-224360e7e887-client-ca\") pod \"route-controller-manager-55d94fcf7b-v9cp9\" (UID: \"23510207-60fc-496b-b907-224360e7e887\") " pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.259653 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23510207-60fc-496b-b907-224360e7e887-serving-cert\") pod \"route-controller-manager-55d94fcf7b-v9cp9\" (UID: \"23510207-60fc-496b-b907-224360e7e887\") " pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.259678 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm5r4\" (UniqueName: \"kubernetes.io/projected/23510207-60fc-496b-b907-224360e7e887-kube-api-access-xm5r4\") pod \"route-controller-manager-55d94fcf7b-v9cp9\" (UID: \"23510207-60fc-496b-b907-224360e7e887\") " pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.260977 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23510207-60fc-496b-b907-224360e7e887-client-ca\") pod \"route-controller-manager-55d94fcf7b-v9cp9\" (UID: \"23510207-60fc-496b-b907-224360e7e887\") " pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.261356 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23510207-60fc-496b-b907-224360e7e887-config\") pod \"route-controller-manager-55d94fcf7b-v9cp9\" (UID: \"23510207-60fc-496b-b907-224360e7e887\") " pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.264801 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23510207-60fc-496b-b907-224360e7e887-serving-cert\") pod \"route-controller-manager-55d94fcf7b-v9cp9\" (UID: \"23510207-60fc-496b-b907-224360e7e887\") " pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.299094 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" event={"ID":"73d513ce-8d01-4a5e-ad2c-a3f86c80668a","Type":"ContainerStarted","Data":"30bedad603a949a671f5e89c22fa86ae81a89a1bed0a3d726dda1b262d0cb66e"} Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.299212 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" event={"ID":"73d513ce-8d01-4a5e-ad2c-a3f86c80668a","Type":"ContainerStarted","Data":"02a01d27feb77bfa412efb4fea128eb7936f5be38a00b8d38940b38f0a2337bc"} Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.299485 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.308148 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm5r4\" (UniqueName: \"kubernetes.io/projected/23510207-60fc-496b-b907-224360e7e887-kube-api-access-xm5r4\") pod \"route-controller-manager-55d94fcf7b-v9cp9\" (UID: \"23510207-60fc-496b-b907-224360e7e887\") " pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.313121 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.323830 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" podStartSLOduration=3.323813145 podStartE2EDuration="3.323813145s" podCreationTimestamp="2026-03-13 09:15:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:15:54.31876163 +0000 UTC m=+195.068676327" watchObservedRunningTime="2026-03-13 09:15:54.323813145 +0000 UTC m=+195.073727822" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.452132 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:54 crc kubenswrapper[4930]: I0313 09:15:54.651619 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9"] Mar 13 09:15:55 crc kubenswrapper[4930]: I0313 09:15:55.307934 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" event={"ID":"23510207-60fc-496b-b907-224360e7e887","Type":"ContainerStarted","Data":"b5434c352bda7c745416492632729de7378d748dea3b07785cf57f04f8ffb95d"} Mar 13 09:15:55 crc kubenswrapper[4930]: I0313 09:15:55.308296 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" event={"ID":"23510207-60fc-496b-b907-224360e7e887","Type":"ContainerStarted","Data":"f17ae5d44e388b8253b84fd3d6e1c4627071bd4ac4cb44dc57c3011f398de542"} Mar 13 09:15:55 crc kubenswrapper[4930]: I0313 09:15:55.330232 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" podStartSLOduration=3.330215266 podStartE2EDuration="3.330215266s" podCreationTimestamp="2026-03-13 09:15:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:15:55.329173658 +0000 UTC m=+196.079088345" watchObservedRunningTime="2026-03-13 09:15:55.330215266 +0000 UTC m=+196.080129943" Mar 13 09:15:56 crc kubenswrapper[4930]: I0313 09:15:56.313945 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:56 crc kubenswrapper[4930]: I0313 09:15:56.321528 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.080238 4930 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.081055 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.081256 4930 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.081669 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://6762958ef4a5f1e05ea149905373f52a6282491fbf69009937361f5bb42b6fc4" gracePeriod=15 Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.081826 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://eb903d30335ef135d24ba2cf2714eb31d42bab12a773363f702b6b416028de80" gracePeriod=15 Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.081858 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://41656d2695e180f85d0efe0f6d6aa8dbed7a3ffd3c3d9cc50c1bdc1f81b81dab" gracePeriod=15 Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.081818 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://459cd20222f850789b336551de6fc7a11362c52d24e9d2ead1faf3531e81e9bd" gracePeriod=15 Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.081824 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://89ff0f4d19a0b5edb6f7c03b5606947f56ff8850698cee5e92ddc35fae32b124" gracePeriod=15 Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.083393 4930 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 09:15:57 crc kubenswrapper[4930]: E0313 09:15:57.083686 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.083712 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 09:15:57 crc kubenswrapper[4930]: E0313 09:15:57.083744 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.083761 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: E0313 09:15:57.083783 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.083798 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 13 09:15:57 crc kubenswrapper[4930]: E0313 09:15:57.083823 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.083839 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 09:15:57 crc kubenswrapper[4930]: E0313 09:15:57.083924 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.083981 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 09:15:57 crc kubenswrapper[4930]: E0313 09:15:57.084048 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.084283 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 09:15:57 crc kubenswrapper[4930]: E0313 09:15:57.084320 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.084339 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: E0313 09:15:57.084364 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.084381 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: E0313 09:15:57.084399 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.084415 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.084713 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.084745 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.084816 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.084835 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.084858 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.084879 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.084901 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: E0313 09:15:57.085127 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.085149 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.085396 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.085429 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.195407 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.195726 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.195756 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.195783 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.195799 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.195821 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.195923 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.196098 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.297768 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.297807 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.297832 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.297848 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.297866 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.297929 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.297936 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.297889 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.297986 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.297985 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.298007 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.298021 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.298056 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.298005 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.298128 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.298205 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.320741 4930 generic.go:334] "Generic (PLEG): container finished" podID="9340acf3-4897-4fb5-8977-61de13adde82" containerID="fcca095ffc1041d2cb023d40521974b76498e8120d0ec351c9dd752604e485a8" exitCode=0 Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.320817 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9340acf3-4897-4fb5-8977-61de13adde82","Type":"ContainerDied","Data":"fcca095ffc1041d2cb023d40521974b76498e8120d0ec351c9dd752604e485a8"} Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.321472 4930 status_manager.go:851] "Failed to get status for pod" podUID="9340acf3-4897-4fb5-8977-61de13adde82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.321805 4930 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.323019 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.324124 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.324666 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="459cd20222f850789b336551de6fc7a11362c52d24e9d2ead1faf3531e81e9bd" exitCode=0 Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.324685 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="41656d2695e180f85d0efe0f6d6aa8dbed7a3ffd3c3d9cc50c1bdc1f81b81dab" exitCode=0 Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.324695 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="eb903d30335ef135d24ba2cf2714eb31d42bab12a773363f702b6b416028de80" exitCode=0 Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.324705 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="89ff0f4d19a0b5edb6f7c03b5606947f56ff8850698cee5e92ddc35fae32b124" exitCode=2 Mar 13 09:15:57 crc kubenswrapper[4930]: I0313 09:15:57.325294 4930 scope.go:117] "RemoveContainer" containerID="1fc8d89dc87d04d45f03cdf66b778aceb7d3a920f5fd254edc0f7d5491c145bf" Mar 13 09:15:58 crc kubenswrapper[4930]: I0313 09:15:58.340976 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 09:15:58 crc kubenswrapper[4930]: I0313 09:15:58.727173 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:58 crc kubenswrapper[4930]: I0313 09:15:58.728610 4930 status_manager.go:851] "Failed to get status for pod" podUID="9340acf3-4897-4fb5-8977-61de13adde82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:15:58 crc kubenswrapper[4930]: I0313 09:15:58.819996 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9340acf3-4897-4fb5-8977-61de13adde82-kubelet-dir\") pod \"9340acf3-4897-4fb5-8977-61de13adde82\" (UID: \"9340acf3-4897-4fb5-8977-61de13adde82\") " Mar 13 09:15:58 crc kubenswrapper[4930]: I0313 09:15:58.820129 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9340acf3-4897-4fb5-8977-61de13adde82-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9340acf3-4897-4fb5-8977-61de13adde82" (UID: "9340acf3-4897-4fb5-8977-61de13adde82"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:15:58 crc kubenswrapper[4930]: I0313 09:15:58.820177 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9340acf3-4897-4fb5-8977-61de13adde82-var-lock\") pod \"9340acf3-4897-4fb5-8977-61de13adde82\" (UID: \"9340acf3-4897-4fb5-8977-61de13adde82\") " Mar 13 09:15:58 crc kubenswrapper[4930]: I0313 09:15:58.820257 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9340acf3-4897-4fb5-8977-61de13adde82-var-lock" (OuterVolumeSpecName: "var-lock") pod "9340acf3-4897-4fb5-8977-61de13adde82" (UID: "9340acf3-4897-4fb5-8977-61de13adde82"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:15:58 crc kubenswrapper[4930]: I0313 09:15:58.820274 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9340acf3-4897-4fb5-8977-61de13adde82-kube-api-access\") pod \"9340acf3-4897-4fb5-8977-61de13adde82\" (UID: \"9340acf3-4897-4fb5-8977-61de13adde82\") " Mar 13 09:15:58 crc kubenswrapper[4930]: I0313 09:15:58.820844 4930 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/9340acf3-4897-4fb5-8977-61de13adde82-var-lock\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:58 crc kubenswrapper[4930]: I0313 09:15:58.820866 4930 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9340acf3-4897-4fb5-8977-61de13adde82-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:58 crc kubenswrapper[4930]: I0313 09:15:58.826652 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9340acf3-4897-4fb5-8977-61de13adde82-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9340acf3-4897-4fb5-8977-61de13adde82" (UID: "9340acf3-4897-4fb5-8977-61de13adde82"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:15:58 crc kubenswrapper[4930]: I0313 09:15:58.922634 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9340acf3-4897-4fb5-8977-61de13adde82-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.348551 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"9340acf3-4897-4fb5-8977-61de13adde82","Type":"ContainerDied","Data":"96d0ff923c0dc44258bf1a8efaaf016429dea5d490d212c4e33b631ed4ddf1c2"} Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.348882 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="96d0ff923c0dc44258bf1a8efaaf016429dea5d490d212c4e33b631ed4ddf1c2" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.348609 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.429340 4930 status_manager.go:851] "Failed to get status for pod" podUID="9340acf3-4897-4fb5-8977-61de13adde82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.432475 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.433354 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.433666 4930 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.433857 4930 status_manager.go:851] "Failed to get status for pod" podUID="9340acf3-4897-4fb5-8977-61de13adde82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.528928 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.529015 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.529108 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.529172 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.529228 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.529418 4930 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.529478 4930 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.529427 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.630425 4930 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.976679 4930 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.977151 4930 status_manager.go:851] "Failed to get status for pod" podUID="9340acf3-4897-4fb5-8977-61de13adde82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:15:59 crc kubenswrapper[4930]: I0313 09:15:59.978885 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.359062 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.359866 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6762958ef4a5f1e05ea149905373f52a6282491fbf69009937361f5bb42b6fc4" exitCode=0 Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.359929 4930 scope.go:117] "RemoveContainer" containerID="459cd20222f850789b336551de6fc7a11362c52d24e9d2ead1faf3531e81e9bd" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.359937 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.361523 4930 status_manager.go:851] "Failed to get status for pod" podUID="9340acf3-4897-4fb5-8977-61de13adde82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.361857 4930 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.363108 4930 status_manager.go:851] "Failed to get status for pod" podUID="9340acf3-4897-4fb5-8977-61de13adde82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.363509 4930 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.378574 4930 scope.go:117] "RemoveContainer" containerID="41656d2695e180f85d0efe0f6d6aa8dbed7a3ffd3c3d9cc50c1bdc1f81b81dab" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.399601 4930 scope.go:117] "RemoveContainer" containerID="eb903d30335ef135d24ba2cf2714eb31d42bab12a773363f702b6b416028de80" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.416084 4930 scope.go:117] "RemoveContainer" containerID="89ff0f4d19a0b5edb6f7c03b5606947f56ff8850698cee5e92ddc35fae32b124" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.429657 4930 scope.go:117] "RemoveContainer" containerID="6762958ef4a5f1e05ea149905373f52a6282491fbf69009937361f5bb42b6fc4" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.445567 4930 scope.go:117] "RemoveContainer" containerID="d564c8034ceace0de6a72404a2167e81b62e93fba97cc5d8ff9b10cf95b16376" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.475464 4930 scope.go:117] "RemoveContainer" containerID="459cd20222f850789b336551de6fc7a11362c52d24e9d2ead1faf3531e81e9bd" Mar 13 09:16:00 crc kubenswrapper[4930]: E0313 09:16:00.475835 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"459cd20222f850789b336551de6fc7a11362c52d24e9d2ead1faf3531e81e9bd\": container with ID starting with 459cd20222f850789b336551de6fc7a11362c52d24e9d2ead1faf3531e81e9bd not found: ID does not exist" containerID="459cd20222f850789b336551de6fc7a11362c52d24e9d2ead1faf3531e81e9bd" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.475868 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"459cd20222f850789b336551de6fc7a11362c52d24e9d2ead1faf3531e81e9bd"} err="failed to get container status \"459cd20222f850789b336551de6fc7a11362c52d24e9d2ead1faf3531e81e9bd\": rpc error: code = NotFound desc = could not find container \"459cd20222f850789b336551de6fc7a11362c52d24e9d2ead1faf3531e81e9bd\": container with ID starting with 459cd20222f850789b336551de6fc7a11362c52d24e9d2ead1faf3531e81e9bd not found: ID does not exist" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.475889 4930 scope.go:117] "RemoveContainer" containerID="41656d2695e180f85d0efe0f6d6aa8dbed7a3ffd3c3d9cc50c1bdc1f81b81dab" Mar 13 09:16:00 crc kubenswrapper[4930]: E0313 09:16:00.476287 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41656d2695e180f85d0efe0f6d6aa8dbed7a3ffd3c3d9cc50c1bdc1f81b81dab\": container with ID starting with 41656d2695e180f85d0efe0f6d6aa8dbed7a3ffd3c3d9cc50c1bdc1f81b81dab not found: ID does not exist" containerID="41656d2695e180f85d0efe0f6d6aa8dbed7a3ffd3c3d9cc50c1bdc1f81b81dab" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.476325 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41656d2695e180f85d0efe0f6d6aa8dbed7a3ffd3c3d9cc50c1bdc1f81b81dab"} err="failed to get container status \"41656d2695e180f85d0efe0f6d6aa8dbed7a3ffd3c3d9cc50c1bdc1f81b81dab\": rpc error: code = NotFound desc = could not find container \"41656d2695e180f85d0efe0f6d6aa8dbed7a3ffd3c3d9cc50c1bdc1f81b81dab\": container with ID starting with 41656d2695e180f85d0efe0f6d6aa8dbed7a3ffd3c3d9cc50c1bdc1f81b81dab not found: ID does not exist" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.476346 4930 scope.go:117] "RemoveContainer" containerID="eb903d30335ef135d24ba2cf2714eb31d42bab12a773363f702b6b416028de80" Mar 13 09:16:00 crc kubenswrapper[4930]: E0313 09:16:00.476768 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb903d30335ef135d24ba2cf2714eb31d42bab12a773363f702b6b416028de80\": container with ID starting with eb903d30335ef135d24ba2cf2714eb31d42bab12a773363f702b6b416028de80 not found: ID does not exist" containerID="eb903d30335ef135d24ba2cf2714eb31d42bab12a773363f702b6b416028de80" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.476808 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb903d30335ef135d24ba2cf2714eb31d42bab12a773363f702b6b416028de80"} err="failed to get container status \"eb903d30335ef135d24ba2cf2714eb31d42bab12a773363f702b6b416028de80\": rpc error: code = NotFound desc = could not find container \"eb903d30335ef135d24ba2cf2714eb31d42bab12a773363f702b6b416028de80\": container with ID starting with eb903d30335ef135d24ba2cf2714eb31d42bab12a773363f702b6b416028de80 not found: ID does not exist" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.476854 4930 scope.go:117] "RemoveContainer" containerID="89ff0f4d19a0b5edb6f7c03b5606947f56ff8850698cee5e92ddc35fae32b124" Mar 13 09:16:00 crc kubenswrapper[4930]: E0313 09:16:00.477181 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89ff0f4d19a0b5edb6f7c03b5606947f56ff8850698cee5e92ddc35fae32b124\": container with ID starting with 89ff0f4d19a0b5edb6f7c03b5606947f56ff8850698cee5e92ddc35fae32b124 not found: ID does not exist" containerID="89ff0f4d19a0b5edb6f7c03b5606947f56ff8850698cee5e92ddc35fae32b124" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.477221 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89ff0f4d19a0b5edb6f7c03b5606947f56ff8850698cee5e92ddc35fae32b124"} err="failed to get container status \"89ff0f4d19a0b5edb6f7c03b5606947f56ff8850698cee5e92ddc35fae32b124\": rpc error: code = NotFound desc = could not find container \"89ff0f4d19a0b5edb6f7c03b5606947f56ff8850698cee5e92ddc35fae32b124\": container with ID starting with 89ff0f4d19a0b5edb6f7c03b5606947f56ff8850698cee5e92ddc35fae32b124 not found: ID does not exist" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.477242 4930 scope.go:117] "RemoveContainer" containerID="6762958ef4a5f1e05ea149905373f52a6282491fbf69009937361f5bb42b6fc4" Mar 13 09:16:00 crc kubenswrapper[4930]: E0313 09:16:00.477739 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6762958ef4a5f1e05ea149905373f52a6282491fbf69009937361f5bb42b6fc4\": container with ID starting with 6762958ef4a5f1e05ea149905373f52a6282491fbf69009937361f5bb42b6fc4 not found: ID does not exist" containerID="6762958ef4a5f1e05ea149905373f52a6282491fbf69009937361f5bb42b6fc4" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.477772 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6762958ef4a5f1e05ea149905373f52a6282491fbf69009937361f5bb42b6fc4"} err="failed to get container status \"6762958ef4a5f1e05ea149905373f52a6282491fbf69009937361f5bb42b6fc4\": rpc error: code = NotFound desc = could not find container \"6762958ef4a5f1e05ea149905373f52a6282491fbf69009937361f5bb42b6fc4\": container with ID starting with 6762958ef4a5f1e05ea149905373f52a6282491fbf69009937361f5bb42b6fc4 not found: ID does not exist" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.477790 4930 scope.go:117] "RemoveContainer" containerID="d564c8034ceace0de6a72404a2167e81b62e93fba97cc5d8ff9b10cf95b16376" Mar 13 09:16:00 crc kubenswrapper[4930]: E0313 09:16:00.478187 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d564c8034ceace0de6a72404a2167e81b62e93fba97cc5d8ff9b10cf95b16376\": container with ID starting with d564c8034ceace0de6a72404a2167e81b62e93fba97cc5d8ff9b10cf95b16376 not found: ID does not exist" containerID="d564c8034ceace0de6a72404a2167e81b62e93fba97cc5d8ff9b10cf95b16376" Mar 13 09:16:00 crc kubenswrapper[4930]: I0313 09:16:00.478233 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d564c8034ceace0de6a72404a2167e81b62e93fba97cc5d8ff9b10cf95b16376"} err="failed to get container status \"d564c8034ceace0de6a72404a2167e81b62e93fba97cc5d8ff9b10cf95b16376\": rpc error: code = NotFound desc = could not find container \"d564c8034ceace0de6a72404a2167e81b62e93fba97cc5d8ff9b10cf95b16376\": container with ID starting with d564c8034ceace0de6a72404a2167e81b62e93fba97cc5d8ff9b10cf95b16376 not found: ID does not exist" Mar 13 09:16:01 crc kubenswrapper[4930]: E0313 09:16:01.133298 4930 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:01 crc kubenswrapper[4930]: E0313 09:16:01.134896 4930 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:01 crc kubenswrapper[4930]: E0313 09:16:01.135510 4930 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:01 crc kubenswrapper[4930]: E0313 09:16:01.136088 4930 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:01 crc kubenswrapper[4930]: E0313 09:16:01.136565 4930 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:01 crc kubenswrapper[4930]: I0313 09:16:01.136633 4930 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 13 09:16:01 crc kubenswrapper[4930]: E0313 09:16:01.137494 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="200ms" Mar 13 09:16:01 crc kubenswrapper[4930]: E0313 09:16:01.339043 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="400ms" Mar 13 09:16:01 crc kubenswrapper[4930]: E0313 09:16:01.740228 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="800ms" Mar 13 09:16:02 crc kubenswrapper[4930]: E0313 09:16:02.115513 4930 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.238:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:16:02 crc kubenswrapper[4930]: I0313 09:16:02.116290 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:16:02 crc kubenswrapper[4930]: W0313 09:16:02.143217 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-c0f0f65c76e47e05927bf2ddad7a61dccda1b7ce6322e825d822732a46904e83 WatchSource:0}: Error finding container c0f0f65c76e47e05927bf2ddad7a61dccda1b7ce6322e825d822732a46904e83: Status 404 returned error can't find the container with id c0f0f65c76e47e05927bf2ddad7a61dccda1b7ce6322e825d822732a46904e83 Mar 13 09:16:02 crc kubenswrapper[4930]: E0313 09:16:02.147152 4930 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.238:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c5bddf5c3dc8b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:16:02.146286731 +0000 UTC m=+202.896201438,LastTimestamp:2026-03-13 09:16:02.146286731 +0000 UTC m=+202.896201438,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:16:02 crc kubenswrapper[4930]: I0313 09:16:02.375357 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"c0f0f65c76e47e05927bf2ddad7a61dccda1b7ce6322e825d822732a46904e83"} Mar 13 09:16:02 crc kubenswrapper[4930]: E0313 09:16:02.543246 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="1.6s" Mar 13 09:16:03 crc kubenswrapper[4930]: I0313 09:16:03.384593 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"8acc57a4ba966c9e35828177fa87a942ca968278af08af1f862037c5456ab39a"} Mar 13 09:16:03 crc kubenswrapper[4930]: I0313 09:16:03.385196 4930 status_manager.go:851] "Failed to get status for pod" podUID="9340acf3-4897-4fb5-8977-61de13adde82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:03 crc kubenswrapper[4930]: E0313 09:16:03.385658 4930 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.238:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:16:04 crc kubenswrapper[4930]: E0313 09:16:04.144241 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="3.2s" Mar 13 09:16:04 crc kubenswrapper[4930]: E0313 09:16:04.392398 4930 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.238:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:16:07 crc kubenswrapper[4930]: E0313 09:16:07.345854 4930 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.238:6443: connect: connection refused" interval="6.4s" Mar 13 09:16:09 crc kubenswrapper[4930]: E0313 09:16:09.727903 4930 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.238:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189c5bddf5c3dc8b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-13 09:16:02.146286731 +0000 UTC m=+202.896201438,LastTimestamp:2026-03-13 09:16:02.146286731 +0000 UTC m=+202.896201438,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 13 09:16:09 crc kubenswrapper[4930]: I0313 09:16:09.975413 4930 status_manager.go:851] "Failed to get status for pod" podUID="9340acf3-4897-4fb5-8977-61de13adde82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:10 crc kubenswrapper[4930]: I0313 09:16:10.970505 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:16:10 crc kubenswrapper[4930]: I0313 09:16:10.971906 4930 status_manager.go:851] "Failed to get status for pod" podUID="9340acf3-4897-4fb5-8977-61de13adde82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:10 crc kubenswrapper[4930]: I0313 09:16:10.988556 4930 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4710e7fe-4c09-4e0e-8e8a-898fddda5e78" Mar 13 09:16:10 crc kubenswrapper[4930]: I0313 09:16:10.988606 4930 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4710e7fe-4c09-4e0e-8e8a-898fddda5e78" Mar 13 09:16:10 crc kubenswrapper[4930]: E0313 09:16:10.989367 4930 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:16:10 crc kubenswrapper[4930]: I0313 09:16:10.990077 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:16:11 crc kubenswrapper[4930]: W0313 09:16:11.019347 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-ff9c3afd65595627919283a19bb224dc8fa179045bd053afd59500e8b0285c80 WatchSource:0}: Error finding container ff9c3afd65595627919283a19bb224dc8fa179045bd053afd59500e8b0285c80: Status 404 returned error can't find the container with id ff9c3afd65595627919283a19bb224dc8fa179045bd053afd59500e8b0285c80 Mar 13 09:16:11 crc kubenswrapper[4930]: I0313 09:16:11.436526 4930 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="4e16a06d8cd599709a3d67c4edbe584cc0806bad4bf9afea6eea0265b376c6e4" exitCode=0 Mar 13 09:16:11 crc kubenswrapper[4930]: I0313 09:16:11.436662 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"4e16a06d8cd599709a3d67c4edbe584cc0806bad4bf9afea6eea0265b376c6e4"} Mar 13 09:16:11 crc kubenswrapper[4930]: I0313 09:16:11.436911 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ff9c3afd65595627919283a19bb224dc8fa179045bd053afd59500e8b0285c80"} Mar 13 09:16:11 crc kubenswrapper[4930]: I0313 09:16:11.437264 4930 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4710e7fe-4c09-4e0e-8e8a-898fddda5e78" Mar 13 09:16:11 crc kubenswrapper[4930]: I0313 09:16:11.437289 4930 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4710e7fe-4c09-4e0e-8e8a-898fddda5e78" Mar 13 09:16:11 crc kubenswrapper[4930]: I0313 09:16:11.437849 4930 status_manager.go:851] "Failed to get status for pod" podUID="9340acf3-4897-4fb5-8977-61de13adde82" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" Mar 13 09:16:11 crc kubenswrapper[4930]: E0313 09:16:11.437867 4930 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.238:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:16:12 crc kubenswrapper[4930]: I0313 09:16:12.309079 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:16:12 crc kubenswrapper[4930]: I0313 09:16:12.309128 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:16:12 crc kubenswrapper[4930]: I0313 09:16:12.451962 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 09:16:12 crc kubenswrapper[4930]: I0313 09:16:12.452674 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 13 09:16:12 crc kubenswrapper[4930]: I0313 09:16:12.452733 4930 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="e3dc535836d240a06323d71572218900609f93fcf435b117c948297f522800cc" exitCode=1 Mar 13 09:16:12 crc kubenswrapper[4930]: I0313 09:16:12.452808 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"e3dc535836d240a06323d71572218900609f93fcf435b117c948297f522800cc"} Mar 13 09:16:12 crc kubenswrapper[4930]: I0313 09:16:12.453491 4930 scope.go:117] "RemoveContainer" containerID="e3dc535836d240a06323d71572218900609f93fcf435b117c948297f522800cc" Mar 13 09:16:12 crc kubenswrapper[4930]: I0313 09:16:12.457947 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b6744c687e6826a22d4422da8f63469cc9f639f1846eb0221175263d682a486b"} Mar 13 09:16:12 crc kubenswrapper[4930]: I0313 09:16:12.457995 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6aff9a3f102b8867cf92a2e7c38e537e70b5e36c484713bbdc5329087b6c993e"} Mar 13 09:16:12 crc kubenswrapper[4930]: I0313 09:16:12.458007 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ddfd060b9e8c7a071bcbe5648cf8fbff87e7a295bb7857c20962229fe8a2f58f"} Mar 13 09:16:13 crc kubenswrapper[4930]: I0313 09:16:13.465116 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 09:16:13 crc kubenswrapper[4930]: I0313 09:16:13.466648 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 13 09:16:13 crc kubenswrapper[4930]: I0313 09:16:13.466729 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"4a1fbef37d7349b1be77763d642a29b52e831bc1c84f43d9151467d8e12a0db8"} Mar 13 09:16:13 crc kubenswrapper[4930]: I0313 09:16:13.469560 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1a17f84d283ffd0952db526b996c555e326b8618e2c4f92c42c784e688733065"} Mar 13 09:16:13 crc kubenswrapper[4930]: I0313 09:16:13.469673 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"4235fe3404595f60fcf35ae2898180419ef94c90baa2108cd2559a80cf9ed6b0"} Mar 13 09:16:13 crc kubenswrapper[4930]: I0313 09:16:13.469746 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:16:13 crc kubenswrapper[4930]: I0313 09:16:13.469902 4930 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4710e7fe-4c09-4e0e-8e8a-898fddda5e78" Mar 13 09:16:13 crc kubenswrapper[4930]: I0313 09:16:13.469934 4930 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4710e7fe-4c09-4e0e-8e8a-898fddda5e78" Mar 13 09:16:15 crc kubenswrapper[4930]: I0313 09:16:15.990729 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:16:15 crc kubenswrapper[4930]: I0313 09:16:15.991494 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:16:16 crc kubenswrapper[4930]: I0313 09:16:16.005962 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:16:16 crc kubenswrapper[4930]: I0313 09:16:16.672576 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:16:16 crc kubenswrapper[4930]: I0313 09:16:16.678988 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:16:17 crc kubenswrapper[4930]: I0313 09:16:17.495030 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:16:18 crc kubenswrapper[4930]: I0313 09:16:18.478103 4930 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:16:18 crc kubenswrapper[4930]: I0313 09:16:18.500341 4930 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4710e7fe-4c09-4e0e-8e8a-898fddda5e78" Mar 13 09:16:18 crc kubenswrapper[4930]: I0313 09:16:18.500380 4930 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4710e7fe-4c09-4e0e-8e8a-898fddda5e78" Mar 13 09:16:18 crc kubenswrapper[4930]: I0313 09:16:18.505982 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:16:19 crc kubenswrapper[4930]: I0313 09:16:19.509417 4930 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4710e7fe-4c09-4e0e-8e8a-898fddda5e78" Mar 13 09:16:19 crc kubenswrapper[4930]: I0313 09:16:19.509489 4930 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4710e7fe-4c09-4e0e-8e8a-898fddda5e78" Mar 13 09:16:20 crc kubenswrapper[4930]: I0313 09:16:19.999990 4930 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="3952d6d7-4045-463a-aea0-36142ce578cc" Mar 13 09:16:28 crc kubenswrapper[4930]: I0313 09:16:28.117333 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 13 09:16:28 crc kubenswrapper[4930]: I0313 09:16:28.195963 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 13 09:16:28 crc kubenswrapper[4930]: I0313 09:16:28.432277 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 13 09:16:28 crc kubenswrapper[4930]: I0313 09:16:28.548087 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 13 09:16:29 crc kubenswrapper[4930]: I0313 09:16:29.274987 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 13 09:16:29 crc kubenswrapper[4930]: I0313 09:16:29.328909 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 09:16:29 crc kubenswrapper[4930]: I0313 09:16:29.332199 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 13 09:16:29 crc kubenswrapper[4930]: I0313 09:16:29.514381 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 13 09:16:29 crc kubenswrapper[4930]: I0313 09:16:29.631996 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 13 09:16:29 crc kubenswrapper[4930]: I0313 09:16:29.663033 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 13 09:16:29 crc kubenswrapper[4930]: I0313 09:16:29.671245 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 13 09:16:29 crc kubenswrapper[4930]: I0313 09:16:29.696357 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 13 09:16:29 crc kubenswrapper[4930]: I0313 09:16:29.887763 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 13 09:16:29 crc kubenswrapper[4930]: I0313 09:16:29.912658 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 13 09:16:30 crc kubenswrapper[4930]: I0313 09:16:30.295337 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 13 09:16:30 crc kubenswrapper[4930]: I0313 09:16:30.318710 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 13 09:16:30 crc kubenswrapper[4930]: I0313 09:16:30.364090 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 13 09:16:30 crc kubenswrapper[4930]: I0313 09:16:30.721200 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 13 09:16:31 crc kubenswrapper[4930]: I0313 09:16:31.008263 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 13 09:16:31 crc kubenswrapper[4930]: I0313 09:16:31.054377 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 13 09:16:31 crc kubenswrapper[4930]: I0313 09:16:31.150430 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 13 09:16:31 crc kubenswrapper[4930]: I0313 09:16:31.187826 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 13 09:16:31 crc kubenswrapper[4930]: I0313 09:16:31.274286 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 13 09:16:31 crc kubenswrapper[4930]: I0313 09:16:31.291408 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 13 09:16:31 crc kubenswrapper[4930]: I0313 09:16:31.403371 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 13 09:16:31 crc kubenswrapper[4930]: I0313 09:16:31.495693 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 13 09:16:31 crc kubenswrapper[4930]: I0313 09:16:31.612806 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 13 09:16:31 crc kubenswrapper[4930]: I0313 09:16:31.663737 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 13 09:16:31 crc kubenswrapper[4930]: I0313 09:16:31.874622 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 13 09:16:31 crc kubenswrapper[4930]: I0313 09:16:31.975216 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 09:16:32 crc kubenswrapper[4930]: I0313 09:16:32.002017 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 13 09:16:32 crc kubenswrapper[4930]: I0313 09:16:32.027116 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 13 09:16:32 crc kubenswrapper[4930]: I0313 09:16:32.076530 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 13 09:16:32 crc kubenswrapper[4930]: I0313 09:16:32.220000 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 13 09:16:32 crc kubenswrapper[4930]: I0313 09:16:32.373808 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 13 09:16:32 crc kubenswrapper[4930]: I0313 09:16:32.531887 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 13 09:16:32 crc kubenswrapper[4930]: I0313 09:16:32.533531 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 13 09:16:32 crc kubenswrapper[4930]: I0313 09:16:32.629625 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 13 09:16:32 crc kubenswrapper[4930]: I0313 09:16:32.819929 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 13 09:16:32 crc kubenswrapper[4930]: I0313 09:16:32.836108 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.041026 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.088013 4930 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.108724 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.189698 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.209680 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.245984 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.344755 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.569889 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.577426 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.609876 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.664337 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.754682 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.762048 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.771209 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.847945 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.890379 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.942247 4930 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.951364 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.962914 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.982565 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.988019 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 13 09:16:33 crc kubenswrapper[4930]: I0313 09:16:33.991518 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.041393 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.255697 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.325340 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.346907 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.430139 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.519933 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.545596 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.655116 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.714612 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.762953 4930 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.768331 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.768413 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.770114 4930 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4710e7fe-4c09-4e0e-8e8a-898fddda5e78" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.770209 4930 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="4710e7fe-4c09-4e0e-8e8a-898fddda5e78" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.790510 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.790493572 podStartE2EDuration="16.790493572s" podCreationTimestamp="2026-03-13 09:16:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:16:34.788039882 +0000 UTC m=+235.537954559" watchObservedRunningTime="2026-03-13 09:16:34.790493572 +0000 UTC m=+235.540408249" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.797310 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.858218 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.861118 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 13 09:16:34 crc kubenswrapper[4930]: I0313 09:16:34.972590 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.024207 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.072328 4930 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.081045 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.085775 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.096365 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.192817 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.213975 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.216393 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.386535 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.425808 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.451336 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.527360 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.573822 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.616140 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.618811 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.660308 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.672740 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.673694 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.723519 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.767655 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.822899 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 13 09:16:35 crc kubenswrapper[4930]: I0313 09:16:35.833220 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.034388 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.063066 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.266703 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.294915 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.304750 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.321656 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.355173 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.429369 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.457430 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.486111 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.486374 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.525678 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.571890 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.621712 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.646678 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.802951 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.878231 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.944739 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 13 09:16:36 crc kubenswrapper[4930]: I0313 09:16:36.965038 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.064490 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.460725 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.468159 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.519668 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.546996 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.826111 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.833845 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.833866 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.864646 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.869544 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.870203 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.891642 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.914375 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.918922 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.977321 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.979997 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 13 09:16:37 crc kubenswrapper[4930]: I0313 09:16:37.995526 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.063163 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.104380 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.200626 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.241276 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.254704 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.284039 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.354474 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.457308 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.534980 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.559551 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.606661 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.641226 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.643889 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.669635 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.700064 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.711728 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.737827 4930 ???:1] "http: TLS handshake error from 192.168.126.11:60554: no serving certificate available for the kubelet" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.778784 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.903358 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.903562 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.910126 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.955506 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 13 09:16:38 crc kubenswrapper[4930]: I0313 09:16:38.981858 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.044749 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.136004 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.140631 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.157191 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.216599 4930 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.235631 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.242124 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.255818 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.287565 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.298143 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.367104 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.385049 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.398512 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.504382 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.555120 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.602080 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.618465 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.816983 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.816983 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.851827 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.881679 4930 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.935452 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 13 09:16:39 crc kubenswrapper[4930]: I0313 09:16:39.937656 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.019801 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.054783 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.140632 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.172365 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.195860 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.204669 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.206938 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.212404 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.218947 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.270471 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.338050 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.358997 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.382563 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.398837 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.406262 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.460849 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556556-58mbz"] Mar 13 09:16:40 crc kubenswrapper[4930]: E0313 09:16:40.461124 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9340acf3-4897-4fb5-8977-61de13adde82" containerName="installer" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.461154 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="9340acf3-4897-4fb5-8977-61de13adde82" containerName="installer" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.461272 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="9340acf3-4897-4fb5-8977-61de13adde82" containerName="installer" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.461925 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556556-58mbz" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.469934 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.470339 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.470678 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.473854 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556556-58mbz"] Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.482781 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7w4g2\" (UniqueName: \"kubernetes.io/projected/146468f7-456c-47bf-b142-cdce5b769152-kube-api-access-7w4g2\") pod \"auto-csr-approver-29556556-58mbz\" (UID: \"146468f7-456c-47bf-b142-cdce5b769152\") " pod="openshift-infra/auto-csr-approver-29556556-58mbz" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.517395 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.555940 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.584136 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7w4g2\" (UniqueName: \"kubernetes.io/projected/146468f7-456c-47bf-b142-cdce5b769152-kube-api-access-7w4g2\") pod \"auto-csr-approver-29556556-58mbz\" (UID: \"146468f7-456c-47bf-b142-cdce5b769152\") " pod="openshift-infra/auto-csr-approver-29556556-58mbz" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.600344 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7w4g2\" (UniqueName: \"kubernetes.io/projected/146468f7-456c-47bf-b142-cdce5b769152-kube-api-access-7w4g2\") pod \"auto-csr-approver-29556556-58mbz\" (UID: \"146468f7-456c-47bf-b142-cdce5b769152\") " pod="openshift-infra/auto-csr-approver-29556556-58mbz" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.630586 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.673222 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.678669 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.754988 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.785999 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556556-58mbz" Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.940017 4930 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 13 09:16:40 crc kubenswrapper[4930]: I0313 09:16:40.940653 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://8acc57a4ba966c9e35828177fa87a942ca968278af08af1f862037c5456ab39a" gracePeriod=5 Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.036726 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.089195 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.097880 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.115254 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.177800 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.237493 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.269799 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556556-58mbz"] Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.309492 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.558033 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.561753 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.576620 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.621056 4930 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.645989 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556556-58mbz" event={"ID":"146468f7-456c-47bf-b142-cdce5b769152","Type":"ContainerStarted","Data":"0fb08b57612ddf0fbf58621d9d17bb7c453d8f3dcfb733013826b201607aefab"} Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.876272 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.892119 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.892939 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.958770 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 13 09:16:41 crc kubenswrapper[4930]: I0313 09:16:41.961929 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.080509 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.135384 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.147886 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.284117 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.308650 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.308951 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.311093 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.337592 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.446970 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.711083 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.769732 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.781540 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 13 09:16:42 crc kubenswrapper[4930]: I0313 09:16:42.818264 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.115702 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.134856 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.135776 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.163544 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.483253 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.497562 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.528697 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.581300 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.592697 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.694408 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.748826 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.888577 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 13 09:16:43 crc kubenswrapper[4930]: I0313 09:16:43.904237 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 13 09:16:44 crc kubenswrapper[4930]: I0313 09:16:44.014297 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 13 09:16:44 crc kubenswrapper[4930]: I0313 09:16:44.238204 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 13 09:16:44 crc kubenswrapper[4930]: I0313 09:16:44.265178 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 13 09:16:44 crc kubenswrapper[4930]: I0313 09:16:44.332012 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 13 09:16:44 crc kubenswrapper[4930]: I0313 09:16:44.333680 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 13 09:16:44 crc kubenswrapper[4930]: I0313 09:16:44.335019 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 13 09:16:44 crc kubenswrapper[4930]: I0313 09:16:44.449491 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 13 09:16:44 crc kubenswrapper[4930]: I0313 09:16:44.504984 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 13 09:16:44 crc kubenswrapper[4930]: I0313 09:16:44.699884 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 13 09:16:44 crc kubenswrapper[4930]: I0313 09:16:44.750344 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 13 09:16:44 crc kubenswrapper[4930]: I0313 09:16:44.854481 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 13 09:16:44 crc kubenswrapper[4930]: I0313 09:16:44.865674 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 13 09:16:45 crc kubenswrapper[4930]: I0313 09:16:45.558797 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 13 09:16:45 crc kubenswrapper[4930]: I0313 09:16:45.849818 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 13 09:16:46 crc kubenswrapper[4930]: I0313 09:16:46.679091 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 09:16:46 crc kubenswrapper[4930]: I0313 09:16:46.679321 4930 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="8acc57a4ba966c9e35828177fa87a942ca968278af08af1f862037c5456ab39a" exitCode=137 Mar 13 09:16:46 crc kubenswrapper[4930]: I0313 09:16:46.776320 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.272057 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.272165 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363202 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363256 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363325 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363332 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363362 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363392 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363444 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363476 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363481 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363730 4930 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363754 4930 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363766 4930 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.363777 4930 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.370458 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.465378 4930 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.684884 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556556-58mbz" event={"ID":"146468f7-456c-47bf-b142-cdce5b769152","Type":"ContainerStarted","Data":"bf57aeac2857ab64e6e2f21e986893a35618c76531f76c7e2f544b8ce79dd1ef"} Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.686304 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.686352 4930 scope.go:117] "RemoveContainer" containerID="8acc57a4ba966c9e35828177fa87a942ca968278af08af1f862037c5456ab39a" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.686493 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.699828 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556556-58mbz" podStartSLOduration=7.633552235 podStartE2EDuration="13.699810425s" podCreationTimestamp="2026-03-13 09:16:34 +0000 UTC" firstStartedPulling="2026-03-13 09:16:41.273014562 +0000 UTC m=+242.022929239" lastFinishedPulling="2026-03-13 09:16:47.339272752 +0000 UTC m=+248.089187429" observedRunningTime="2026-03-13 09:16:47.699241151 +0000 UTC m=+248.449155828" watchObservedRunningTime="2026-03-13 09:16:47.699810425 +0000 UTC m=+248.449725102" Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.851270 4930 csr.go:261] certificate signing request csr-clgbp is approved, waiting to be issued Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.870041 4930 csr.go:257] certificate signing request csr-clgbp is issued Mar 13 09:16:47 crc kubenswrapper[4930]: I0313 09:16:47.977148 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 13 09:16:48 crc kubenswrapper[4930]: I0313 09:16:48.692270 4930 generic.go:334] "Generic (PLEG): container finished" podID="146468f7-456c-47bf-b142-cdce5b769152" containerID="bf57aeac2857ab64e6e2f21e986893a35618c76531f76c7e2f544b8ce79dd1ef" exitCode=0 Mar 13 09:16:48 crc kubenswrapper[4930]: I0313 09:16:48.692351 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556556-58mbz" event={"ID":"146468f7-456c-47bf-b142-cdce5b769152","Type":"ContainerDied","Data":"bf57aeac2857ab64e6e2f21e986893a35618c76531f76c7e2f544b8ce79dd1ef"} Mar 13 09:16:48 crc kubenswrapper[4930]: I0313 09:16:48.870986 4930 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-13 10:37:08.838242739 +0000 UTC Mar 13 09:16:48 crc kubenswrapper[4930]: I0313 09:16:48.871022 4930 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 5881h20m19.967224105s for next certificate rotation Mar 13 09:16:49 crc kubenswrapper[4930]: I0313 09:16:49.871817 4930 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-31 15:18:11.361557173 +0000 UTC Mar 13 09:16:49 crc kubenswrapper[4930]: I0313 09:16:49.871852 4930 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 7038h1m21.48970743s for next certificate rotation Mar 13 09:16:49 crc kubenswrapper[4930]: I0313 09:16:49.976602 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556556-58mbz" Mar 13 09:16:49 crc kubenswrapper[4930]: I0313 09:16:49.994655 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7w4g2\" (UniqueName: \"kubernetes.io/projected/146468f7-456c-47bf-b142-cdce5b769152-kube-api-access-7w4g2\") pod \"146468f7-456c-47bf-b142-cdce5b769152\" (UID: \"146468f7-456c-47bf-b142-cdce5b769152\") " Mar 13 09:16:50 crc kubenswrapper[4930]: I0313 09:16:50.000833 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/146468f7-456c-47bf-b142-cdce5b769152-kube-api-access-7w4g2" (OuterVolumeSpecName: "kube-api-access-7w4g2") pod "146468f7-456c-47bf-b142-cdce5b769152" (UID: "146468f7-456c-47bf-b142-cdce5b769152"). InnerVolumeSpecName "kube-api-access-7w4g2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:16:50 crc kubenswrapper[4930]: I0313 09:16:50.096957 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7w4g2\" (UniqueName: \"kubernetes.io/projected/146468f7-456c-47bf-b142-cdce5b769152-kube-api-access-7w4g2\") on node \"crc\" DevicePath \"\"" Mar 13 09:16:50 crc kubenswrapper[4930]: I0313 09:16:50.702642 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556556-58mbz" event={"ID":"146468f7-456c-47bf-b142-cdce5b769152","Type":"ContainerDied","Data":"0fb08b57612ddf0fbf58621d9d17bb7c453d8f3dcfb733013826b201607aefab"} Mar 13 09:16:50 crc kubenswrapper[4930]: I0313 09:16:50.702705 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fb08b57612ddf0fbf58621d9d17bb7c453d8f3dcfb733013826b201607aefab" Mar 13 09:16:50 crc kubenswrapper[4930]: I0313 09:16:50.702702 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556556-58mbz" Mar 13 09:17:04 crc kubenswrapper[4930]: I0313 09:17:04.782327 4930 generic.go:334] "Generic (PLEG): container finished" podID="bb104f34-0d73-40d9-9ba2-13401d61315d" containerID="cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe" exitCode=0 Mar 13 09:17:04 crc kubenswrapper[4930]: I0313 09:17:04.782477 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" event={"ID":"bb104f34-0d73-40d9-9ba2-13401d61315d","Type":"ContainerDied","Data":"cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe"} Mar 13 09:17:04 crc kubenswrapper[4930]: I0313 09:17:04.783467 4930 scope.go:117] "RemoveContainer" containerID="cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe" Mar 13 09:17:05 crc kubenswrapper[4930]: I0313 09:17:05.790983 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" event={"ID":"bb104f34-0d73-40d9-9ba2-13401d61315d","Type":"ContainerStarted","Data":"0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028"} Mar 13 09:17:05 crc kubenswrapper[4930]: I0313 09:17:05.792791 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:17:05 crc kubenswrapper[4930]: I0313 09:17:05.794142 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:17:12 crc kubenswrapper[4930]: I0313 09:17:12.308991 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:17:12 crc kubenswrapper[4930]: I0313 09:17:12.309858 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:17:12 crc kubenswrapper[4930]: I0313 09:17:12.309929 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:17:12 crc kubenswrapper[4930]: I0313 09:17:12.310689 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"75a32c3a6a592ba94cf2f17752d6cd696cc0c651141c5d595391ac31dc2f4185"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 09:17:12 crc kubenswrapper[4930]: I0313 09:17:12.310786 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://75a32c3a6a592ba94cf2f17752d6cd696cc0c651141c5d595391ac31dc2f4185" gracePeriod=600 Mar 13 09:17:12 crc kubenswrapper[4930]: I0313 09:17:12.832762 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="75a32c3a6a592ba94cf2f17752d6cd696cc0c651141c5d595391ac31dc2f4185" exitCode=0 Mar 13 09:17:12 crc kubenswrapper[4930]: I0313 09:17:12.832845 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"75a32c3a6a592ba94cf2f17752d6cd696cc0c651141c5d595391ac31dc2f4185"} Mar 13 09:17:12 crc kubenswrapper[4930]: I0313 09:17:12.833048 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"df84b7067d27878debd4bec4f38e693600039c3ae7e13f91f3d5b6a67bef6a54"} Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.439592 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4sq9d"] Mar 13 09:17:51 crc kubenswrapper[4930]: E0313 09:17:51.440393 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.440406 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 09:17:51 crc kubenswrapper[4930]: E0313 09:17:51.440416 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="146468f7-456c-47bf-b142-cdce5b769152" containerName="oc" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.440423 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="146468f7-456c-47bf-b142-cdce5b769152" containerName="oc" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.440627 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.440646 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="146468f7-456c-47bf-b142-cdce5b769152" containerName="oc" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.441265 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.451913 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4sq9d"] Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.535236 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9027107f-0c82-4dd0-a8f1-01d09d646077-trusted-ca\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.535308 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9027107f-0c82-4dd0-a8f1-01d09d646077-bound-sa-token\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.535346 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.535406 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8b4h\" (UniqueName: \"kubernetes.io/projected/9027107f-0c82-4dd0-a8f1-01d09d646077-kube-api-access-l8b4h\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.535453 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9027107f-0c82-4dd0-a8f1-01d09d646077-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.535638 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9027107f-0c82-4dd0-a8f1-01d09d646077-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.535698 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9027107f-0c82-4dd0-a8f1-01d09d646077-registry-certificates\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.535797 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9027107f-0c82-4dd0-a8f1-01d09d646077-registry-tls\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.558888 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.636601 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9027107f-0c82-4dd0-a8f1-01d09d646077-registry-tls\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.636658 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9027107f-0c82-4dd0-a8f1-01d09d646077-trusted-ca\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.636706 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9027107f-0c82-4dd0-a8f1-01d09d646077-bound-sa-token\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.636737 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8b4h\" (UniqueName: \"kubernetes.io/projected/9027107f-0c82-4dd0-a8f1-01d09d646077-kube-api-access-l8b4h\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.636762 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9027107f-0c82-4dd0-a8f1-01d09d646077-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.636788 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9027107f-0c82-4dd0-a8f1-01d09d646077-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.636803 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9027107f-0c82-4dd0-a8f1-01d09d646077-registry-certificates\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.637562 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/9027107f-0c82-4dd0-a8f1-01d09d646077-ca-trust-extracted\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.638027 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/9027107f-0c82-4dd0-a8f1-01d09d646077-registry-certificates\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.638898 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9027107f-0c82-4dd0-a8f1-01d09d646077-trusted-ca\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.642110 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/9027107f-0c82-4dd0-a8f1-01d09d646077-installation-pull-secrets\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.642317 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/9027107f-0c82-4dd0-a8f1-01d09d646077-registry-tls\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.652214 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/9027107f-0c82-4dd0-a8f1-01d09d646077-bound-sa-token\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.654630 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8b4h\" (UniqueName: \"kubernetes.io/projected/9027107f-0c82-4dd0-a8f1-01d09d646077-kube-api-access-l8b4h\") pod \"image-registry-66df7c8f76-4sq9d\" (UID: \"9027107f-0c82-4dd0-a8f1-01d09d646077\") " pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:51 crc kubenswrapper[4930]: I0313 09:17:51.758919 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:52 crc kubenswrapper[4930]: I0313 09:17:52.178511 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-4sq9d"] Mar 13 09:17:53 crc kubenswrapper[4930]: I0313 09:17:53.064126 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" event={"ID":"9027107f-0c82-4dd0-a8f1-01d09d646077","Type":"ContainerStarted","Data":"44f5e4ecf8f4e5ee48a38b732d2f54373eacdaaf3ce87c4d98f79cbb6f80ba2f"} Mar 13 09:17:53 crc kubenswrapper[4930]: I0313 09:17:53.065647 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:17:53 crc kubenswrapper[4930]: I0313 09:17:53.065725 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" event={"ID":"9027107f-0c82-4dd0-a8f1-01d09d646077","Type":"ContainerStarted","Data":"32b1a35ad7d0a88339dc7df7d4320fc9094552da077a20628bec2dd69d1fdce2"} Mar 13 09:17:53 crc kubenswrapper[4930]: I0313 09:17:53.088773 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" podStartSLOduration=2.088753463 podStartE2EDuration="2.088753463s" podCreationTimestamp="2026-03-13 09:17:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:17:53.081973865 +0000 UTC m=+313.831888582" watchObservedRunningTime="2026-03-13 09:17:53.088753463 +0000 UTC m=+313.838668150" Mar 13 09:18:00 crc kubenswrapper[4930]: I0313 09:18:00.133745 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556558-j275q"] Mar 13 09:18:00 crc kubenswrapper[4930]: I0313 09:18:00.135356 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556558-j275q" Mar 13 09:18:00 crc kubenswrapper[4930]: I0313 09:18:00.138148 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:18:00 crc kubenswrapper[4930]: I0313 09:18:00.138352 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:18:00 crc kubenswrapper[4930]: I0313 09:18:00.138374 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:18:00 crc kubenswrapper[4930]: I0313 09:18:00.141600 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556558-j275q"] Mar 13 09:18:00 crc kubenswrapper[4930]: I0313 09:18:00.231085 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq7w2\" (UniqueName: \"kubernetes.io/projected/07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1-kube-api-access-sq7w2\") pod \"auto-csr-approver-29556558-j275q\" (UID: \"07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1\") " pod="openshift-infra/auto-csr-approver-29556558-j275q" Mar 13 09:18:00 crc kubenswrapper[4930]: I0313 09:18:00.332491 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq7w2\" (UniqueName: \"kubernetes.io/projected/07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1-kube-api-access-sq7w2\") pod \"auto-csr-approver-29556558-j275q\" (UID: \"07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1\") " pod="openshift-infra/auto-csr-approver-29556558-j275q" Mar 13 09:18:00 crc kubenswrapper[4930]: I0313 09:18:00.352162 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq7w2\" (UniqueName: \"kubernetes.io/projected/07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1-kube-api-access-sq7w2\") pod \"auto-csr-approver-29556558-j275q\" (UID: \"07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1\") " pod="openshift-infra/auto-csr-approver-29556558-j275q" Mar 13 09:18:00 crc kubenswrapper[4930]: I0313 09:18:00.453157 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556558-j275q" Mar 13 09:18:00 crc kubenswrapper[4930]: I0313 09:18:00.880057 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556558-j275q"] Mar 13 09:18:01 crc kubenswrapper[4930]: I0313 09:18:01.109888 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556558-j275q" event={"ID":"07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1","Type":"ContainerStarted","Data":"72c9d9099c5188ce4f584942c3e3e0d8ce7101d0b8d9ace7d9a9a47bb60d4831"} Mar 13 09:18:03 crc kubenswrapper[4930]: I0313 09:18:03.120111 4930 generic.go:334] "Generic (PLEG): container finished" podID="07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1" containerID="9bd519de18d33d4bec56a1f3c7b725899af16bae93489a1d64e20c746cd894d1" exitCode=0 Mar 13 09:18:03 crc kubenswrapper[4930]: I0313 09:18:03.120706 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556558-j275q" event={"ID":"07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1","Type":"ContainerDied","Data":"9bd519de18d33d4bec56a1f3c7b725899af16bae93489a1d64e20c746cd894d1"} Mar 13 09:18:04 crc kubenswrapper[4930]: I0313 09:18:04.332179 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556558-j275q" Mar 13 09:18:04 crc kubenswrapper[4930]: I0313 09:18:04.409503 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq7w2\" (UniqueName: \"kubernetes.io/projected/07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1-kube-api-access-sq7w2\") pod \"07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1\" (UID: \"07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1\") " Mar 13 09:18:04 crc kubenswrapper[4930]: I0313 09:18:04.414401 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1-kube-api-access-sq7w2" (OuterVolumeSpecName: "kube-api-access-sq7w2") pod "07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1" (UID: "07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1"). InnerVolumeSpecName "kube-api-access-sq7w2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:18:04 crc kubenswrapper[4930]: I0313 09:18:04.510361 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq7w2\" (UniqueName: \"kubernetes.io/projected/07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1-kube-api-access-sq7w2\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:05 crc kubenswrapper[4930]: I0313 09:18:05.135048 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556558-j275q" event={"ID":"07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1","Type":"ContainerDied","Data":"72c9d9099c5188ce4f584942c3e3e0d8ce7101d0b8d9ace7d9a9a47bb60d4831"} Mar 13 09:18:05 crc kubenswrapper[4930]: I0313 09:18:05.135288 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="72c9d9099c5188ce4f584942c3e3e0d8ce7101d0b8d9ace7d9a9a47bb60d4831" Mar 13 09:18:05 crc kubenswrapper[4930]: I0313 09:18:05.135123 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556558-j275q" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.001130 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tnq4n"] Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.001727 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tnq4n" podUID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" containerName="registry-server" containerID="cri-o://dcf841d84dca2fde34daf6d3dd2b33890ac1c4ae49782c2d6977ba7a83f68c5b" gracePeriod=30 Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.041163 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zrbn4"] Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.042452 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-zrbn4" podUID="bf6e93be-455f-45e4-a51b-d851591c31ed" containerName="registry-server" containerID="cri-o://f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e" gracePeriod=30 Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.056958 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdbqj"] Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.057345 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" podUID="bb104f34-0d73-40d9-9ba2-13401d61315d" containerName="marketplace-operator" containerID="cri-o://0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028" gracePeriod=30 Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.060171 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b56p5"] Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.060375 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b56p5" podUID="085d31a1-d613-4e2f-a992-42bfc9b66413" containerName="registry-server" containerID="cri-o://f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd" gracePeriod=30 Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.068688 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zptn9"] Mar 13 09:18:06 crc kubenswrapper[4930]: E0313 09:18:06.068898 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1" containerName="oc" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.068909 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1" containerName="oc" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.069013 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1" containerName="oc" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.069637 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.073219 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cvrjq"] Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.073464 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cvrjq" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerName="registry-server" containerID="cri-o://79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2" gracePeriod=30 Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.078841 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zptn9"] Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.132418 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d705c5d3-c140-479e-962c-958a2366753c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zptn9\" (UID: \"d705c5d3-c140-479e-962c-958a2366753c\") " pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.132539 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv29x\" (UniqueName: \"kubernetes.io/projected/d705c5d3-c140-479e-962c-958a2366753c-kube-api-access-zv29x\") pod \"marketplace-operator-79b997595-zptn9\" (UID: \"d705c5d3-c140-479e-962c-958a2366753c\") " pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.132568 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d705c5d3-c140-479e-962c-958a2366753c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zptn9\" (UID: \"d705c5d3-c140-479e-962c-958a2366753c\") " pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.142190 4930 generic.go:334] "Generic (PLEG): container finished" podID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" containerID="dcf841d84dca2fde34daf6d3dd2b33890ac1c4ae49782c2d6977ba7a83f68c5b" exitCode=0 Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.142236 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tnq4n" event={"ID":"342671c3-b1f8-4c61-a0da-abe7dcdfc367","Type":"ContainerDied","Data":"dcf841d84dca2fde34daf6d3dd2b33890ac1c4ae49782c2d6977ba7a83f68c5b"} Mar 13 09:18:06 crc kubenswrapper[4930]: E0313 09:18:06.181486 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2 is running failed: container process not found" containerID="79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 09:18:06 crc kubenswrapper[4930]: E0313 09:18:06.181763 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2 is running failed: container process not found" containerID="79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 09:18:06 crc kubenswrapper[4930]: E0313 09:18:06.184476 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2 is running failed: container process not found" containerID="79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 09:18:06 crc kubenswrapper[4930]: E0313 09:18:06.184518 4930 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-cvrjq" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerName="registry-server" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.234055 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv29x\" (UniqueName: \"kubernetes.io/projected/d705c5d3-c140-479e-962c-958a2366753c-kube-api-access-zv29x\") pod \"marketplace-operator-79b997595-zptn9\" (UID: \"d705c5d3-c140-479e-962c-958a2366753c\") " pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.234101 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d705c5d3-c140-479e-962c-958a2366753c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zptn9\" (UID: \"d705c5d3-c140-479e-962c-958a2366753c\") " pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.234193 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d705c5d3-c140-479e-962c-958a2366753c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zptn9\" (UID: \"d705c5d3-c140-479e-962c-958a2366753c\") " pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.235454 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d705c5d3-c140-479e-962c-958a2366753c-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zptn9\" (UID: \"d705c5d3-c140-479e-962c-958a2366753c\") " pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.248910 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d705c5d3-c140-479e-962c-958a2366753c-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zptn9\" (UID: \"d705c5d3-c140-479e-962c-958a2366753c\") " pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.251483 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv29x\" (UniqueName: \"kubernetes.io/projected/d705c5d3-c140-479e-962c-958a2366753c-kube-api-access-zv29x\") pod \"marketplace-operator-79b997595-zptn9\" (UID: \"d705c5d3-c140-479e-962c-958a2366753c\") " pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.426645 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.432942 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.436673 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zqw8x\" (UniqueName: \"kubernetes.io/projected/342671c3-b1f8-4c61-a0da-abe7dcdfc367-kube-api-access-zqw8x\") pod \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\" (UID: \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.436725 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/342671c3-b1f8-4c61-a0da-abe7dcdfc367-utilities\") pod \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\" (UID: \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.436813 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/342671c3-b1f8-4c61-a0da-abe7dcdfc367-catalog-content\") pod \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\" (UID: \"342671c3-b1f8-4c61-a0da-abe7dcdfc367\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.437651 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/342671c3-b1f8-4c61-a0da-abe7dcdfc367-utilities" (OuterVolumeSpecName: "utilities") pod "342671c3-b1f8-4c61-a0da-abe7dcdfc367" (UID: "342671c3-b1f8-4c61-a0da-abe7dcdfc367"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.447384 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/342671c3-b1f8-4c61-a0da-abe7dcdfc367-kube-api-access-zqw8x" (OuterVolumeSpecName: "kube-api-access-zqw8x") pod "342671c3-b1f8-4c61-a0da-abe7dcdfc367" (UID: "342671c3-b1f8-4c61-a0da-abe7dcdfc367"). InnerVolumeSpecName "kube-api-access-zqw8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.463677 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.490217 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.494547 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.503303 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.540679 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb104f34-0d73-40d9-9ba2-13401d61315d-marketplace-trusted-ca\") pod \"bb104f34-0d73-40d9-9ba2-13401d61315d\" (UID: \"bb104f34-0d73-40d9-9ba2-13401d61315d\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.540734 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6e93be-455f-45e4-a51b-d851591c31ed-catalog-content\") pod \"bf6e93be-455f-45e4-a51b-d851591c31ed\" (UID: \"bf6e93be-455f-45e4-a51b-d851591c31ed\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.540799 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k7ln6\" (UniqueName: \"kubernetes.io/projected/bf6e93be-455f-45e4-a51b-d851591c31ed-kube-api-access-k7ln6\") pod \"bf6e93be-455f-45e4-a51b-d851591c31ed\" (UID: \"bf6e93be-455f-45e4-a51b-d851591c31ed\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.540824 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-catalog-content\") pod \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\" (UID: \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.540852 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9bxr\" (UniqueName: \"kubernetes.io/projected/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-kube-api-access-z9bxr\") pod \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\" (UID: \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.540887 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/085d31a1-d613-4e2f-a992-42bfc9b66413-utilities\") pod \"085d31a1-d613-4e2f-a992-42bfc9b66413\" (UID: \"085d31a1-d613-4e2f-a992-42bfc9b66413\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.540906 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7jl4n\" (UniqueName: \"kubernetes.io/projected/085d31a1-d613-4e2f-a992-42bfc9b66413-kube-api-access-7jl4n\") pod \"085d31a1-d613-4e2f-a992-42bfc9b66413\" (UID: \"085d31a1-d613-4e2f-a992-42bfc9b66413\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.540963 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-utilities\") pod \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\" (UID: \"5933708c-4e7f-4567-bd1e-df3dfe85fcd5\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.540984 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-stx2q\" (UniqueName: \"kubernetes.io/projected/bb104f34-0d73-40d9-9ba2-13401d61315d-kube-api-access-stx2q\") pod \"bb104f34-0d73-40d9-9ba2-13401d61315d\" (UID: \"bb104f34-0d73-40d9-9ba2-13401d61315d\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.541000 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6e93be-455f-45e4-a51b-d851591c31ed-utilities\") pod \"bf6e93be-455f-45e4-a51b-d851591c31ed\" (UID: \"bf6e93be-455f-45e4-a51b-d851591c31ed\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.541029 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb104f34-0d73-40d9-9ba2-13401d61315d-marketplace-operator-metrics\") pod \"bb104f34-0d73-40d9-9ba2-13401d61315d\" (UID: \"bb104f34-0d73-40d9-9ba2-13401d61315d\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.541051 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/085d31a1-d613-4e2f-a992-42bfc9b66413-catalog-content\") pod \"085d31a1-d613-4e2f-a992-42bfc9b66413\" (UID: \"085d31a1-d613-4e2f-a992-42bfc9b66413\") " Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.541322 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zqw8x\" (UniqueName: \"kubernetes.io/projected/342671c3-b1f8-4c61-a0da-abe7dcdfc367-kube-api-access-zqw8x\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.541333 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/342671c3-b1f8-4c61-a0da-abe7dcdfc367-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.544164 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bb104f34-0d73-40d9-9ba2-13401d61315d-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "bb104f34-0d73-40d9-9ba2-13401d61315d" (UID: "bb104f34-0d73-40d9-9ba2-13401d61315d"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.550771 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf6e93be-455f-45e4-a51b-d851591c31ed-utilities" (OuterVolumeSpecName: "utilities") pod "bf6e93be-455f-45e4-a51b-d851591c31ed" (UID: "bf6e93be-455f-45e4-a51b-d851591c31ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.551385 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-utilities" (OuterVolumeSpecName: "utilities") pod "5933708c-4e7f-4567-bd1e-df3dfe85fcd5" (UID: "5933708c-4e7f-4567-bd1e-df3dfe85fcd5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.551606 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/085d31a1-d613-4e2f-a992-42bfc9b66413-utilities" (OuterVolumeSpecName: "utilities") pod "085d31a1-d613-4e2f-a992-42bfc9b66413" (UID: "085d31a1-d613-4e2f-a992-42bfc9b66413"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.552809 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/085d31a1-d613-4e2f-a992-42bfc9b66413-kube-api-access-7jl4n" (OuterVolumeSpecName: "kube-api-access-7jl4n") pod "085d31a1-d613-4e2f-a992-42bfc9b66413" (UID: "085d31a1-d613-4e2f-a992-42bfc9b66413"). InnerVolumeSpecName "kube-api-access-7jl4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.553330 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-kube-api-access-z9bxr" (OuterVolumeSpecName: "kube-api-access-z9bxr") pod "5933708c-4e7f-4567-bd1e-df3dfe85fcd5" (UID: "5933708c-4e7f-4567-bd1e-df3dfe85fcd5"). InnerVolumeSpecName "kube-api-access-z9bxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.553441 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb104f34-0d73-40d9-9ba2-13401d61315d-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "bb104f34-0d73-40d9-9ba2-13401d61315d" (UID: "bb104f34-0d73-40d9-9ba2-13401d61315d"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.555803 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb104f34-0d73-40d9-9ba2-13401d61315d-kube-api-access-stx2q" (OuterVolumeSpecName: "kube-api-access-stx2q") pod "bb104f34-0d73-40d9-9ba2-13401d61315d" (UID: "bb104f34-0d73-40d9-9ba2-13401d61315d"). InnerVolumeSpecName "kube-api-access-stx2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.556552 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf6e93be-455f-45e4-a51b-d851591c31ed-kube-api-access-k7ln6" (OuterVolumeSpecName: "kube-api-access-k7ln6") pod "bf6e93be-455f-45e4-a51b-d851591c31ed" (UID: "bf6e93be-455f-45e4-a51b-d851591c31ed"). InnerVolumeSpecName "kube-api-access-k7ln6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.571601 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/342671c3-b1f8-4c61-a0da-abe7dcdfc367-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "342671c3-b1f8-4c61-a0da-abe7dcdfc367" (UID: "342671c3-b1f8-4c61-a0da-abe7dcdfc367"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.598978 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/085d31a1-d613-4e2f-a992-42bfc9b66413-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "085d31a1-d613-4e2f-a992-42bfc9b66413" (UID: "085d31a1-d613-4e2f-a992-42bfc9b66413"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.615609 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bf6e93be-455f-45e4-a51b-d851591c31ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bf6e93be-455f-45e4-a51b-d851591c31ed" (UID: "bf6e93be-455f-45e4-a51b-d851591c31ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.642083 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9bxr\" (UniqueName: \"kubernetes.io/projected/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-kube-api-access-z9bxr\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.642114 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/085d31a1-d613-4e2f-a992-42bfc9b66413-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.642124 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7jl4n\" (UniqueName: \"kubernetes.io/projected/085d31a1-d613-4e2f-a992-42bfc9b66413-kube-api-access-7jl4n\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.642135 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.642144 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-stx2q\" (UniqueName: \"kubernetes.io/projected/bb104f34-0d73-40d9-9ba2-13401d61315d-kube-api-access-stx2q\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.642152 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bf6e93be-455f-45e4-a51b-d851591c31ed-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.642160 4930 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/bb104f34-0d73-40d9-9ba2-13401d61315d-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.642169 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/085d31a1-d613-4e2f-a992-42bfc9b66413-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.642178 4930 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bb104f34-0d73-40d9-9ba2-13401d61315d-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.642186 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bf6e93be-455f-45e4-a51b-d851591c31ed-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.642196 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/342671c3-b1f8-4c61-a0da-abe7dcdfc367-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.642204 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k7ln6\" (UniqueName: \"kubernetes.io/projected/bf6e93be-455f-45e4-a51b-d851591c31ed-kube-api-access-k7ln6\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.707883 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5933708c-4e7f-4567-bd1e-df3dfe85fcd5" (UID: "5933708c-4e7f-4567-bd1e-df3dfe85fcd5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.743233 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5933708c-4e7f-4567-bd1e-df3dfe85fcd5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:06 crc kubenswrapper[4930]: I0313 09:18:06.871026 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zptn9"] Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.148357 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" event={"ID":"d705c5d3-c140-479e-962c-958a2366753c","Type":"ContainerStarted","Data":"ef3feae5d76af1224b6c5c69d58de1290c3714a1add90314bbd94652c51a7da9"} Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.148723 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" event={"ID":"d705c5d3-c140-479e-962c-958a2366753c","Type":"ContainerStarted","Data":"6e2d34f63a74041ab626c87c380f20647b14a471aa1ce69fdb769b6d44831e8d"} Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.148750 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.149809 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zptn9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" start-of-body= Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.149863 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.153449 4930 generic.go:334] "Generic (PLEG): container finished" podID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerID="79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2" exitCode=0 Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.153513 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvrjq" event={"ID":"5933708c-4e7f-4567-bd1e-df3dfe85fcd5","Type":"ContainerDied","Data":"79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2"} Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.153542 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cvrjq" event={"ID":"5933708c-4e7f-4567-bd1e-df3dfe85fcd5","Type":"ContainerDied","Data":"3b718bc407c8d2708b66369e8b498b225b48037371e4cc333f9029d9cc83be8a"} Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.153559 4930 scope.go:117] "RemoveContainer" containerID="79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.153653 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cvrjq" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.157705 4930 generic.go:334] "Generic (PLEG): container finished" podID="bb104f34-0d73-40d9-9ba2-13401d61315d" containerID="0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028" exitCode=0 Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.157728 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.157808 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" event={"ID":"bb104f34-0d73-40d9-9ba2-13401d61315d","Type":"ContainerDied","Data":"0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028"} Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.157837 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-rdbqj" event={"ID":"bb104f34-0d73-40d9-9ba2-13401d61315d","Type":"ContainerDied","Data":"b59038704e56512ca177e9a750f3c1e6851fc8da6025829e26d83d6b7b07a049"} Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.160558 4930 generic.go:334] "Generic (PLEG): container finished" podID="bf6e93be-455f-45e4-a51b-d851591c31ed" containerID="f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e" exitCode=0 Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.160668 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrbn4" event={"ID":"bf6e93be-455f-45e4-a51b-d851591c31ed","Type":"ContainerDied","Data":"f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e"} Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.160717 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-zrbn4" event={"ID":"bf6e93be-455f-45e4-a51b-d851591c31ed","Type":"ContainerDied","Data":"810db0484195e739096e8ec2db9f56086c330bb79c5dbd369184adfc6fc6e9bb"} Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.160822 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-zrbn4" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.164362 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tnq4n" event={"ID":"342671c3-b1f8-4c61-a0da-abe7dcdfc367","Type":"ContainerDied","Data":"55edb598698e63a1ac98360e709600815b6e72f8d2b5c6e8e6552c1a07631568"} Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.164473 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tnq4n" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.170402 4930 generic.go:334] "Generic (PLEG): container finished" podID="085d31a1-d613-4e2f-a992-42bfc9b66413" containerID="f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd" exitCode=0 Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.170462 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b56p5" event={"ID":"085d31a1-d613-4e2f-a992-42bfc9b66413","Type":"ContainerDied","Data":"f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd"} Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.170487 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b56p5" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.170493 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b56p5" event={"ID":"085d31a1-d613-4e2f-a992-42bfc9b66413","Type":"ContainerDied","Data":"0a9faadb42f5ab5c4ffdabf51f9741cf4ef9e2b5bde2eceebcb82176ded4e5cc"} Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.172491 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podStartSLOduration=1.172478642 podStartE2EDuration="1.172478642s" podCreationTimestamp="2026-03-13 09:18:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:18:07.167031577 +0000 UTC m=+327.916946254" watchObservedRunningTime="2026-03-13 09:18:07.172478642 +0000 UTC m=+327.922393319" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.211165 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdbqj"] Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.215136 4930 scope.go:117] "RemoveContainer" containerID="cddfdd5aa1f76cb6428df177e8b597ff13e0216ebc2def1eed47177c18b7b8e5" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.217710 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-rdbqj"] Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.223158 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b56p5"] Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.225913 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b56p5"] Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.232877 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-zrbn4"] Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.238760 4930 scope.go:117] "RemoveContainer" containerID="7ee210e26eaadde01e83f39b637dc5c600bd87d87402fcc3f3a29904b1b03c34" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.244742 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-zrbn4"] Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.253642 4930 scope.go:117] "RemoveContainer" containerID="79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2" Mar 13 09:18:07 crc kubenswrapper[4930]: E0313 09:18:07.264148 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2\": container with ID starting with 79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2 not found: ID does not exist" containerID="79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.264195 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2"} err="failed to get container status \"79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2\": rpc error: code = NotFound desc = could not find container \"79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2\": container with ID starting with 79a763acd1abb90b71a9ff1491dc90de2c204a54ae5e88158036d86f2683bff2 not found: ID does not exist" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.264220 4930 scope.go:117] "RemoveContainer" containerID="cddfdd5aa1f76cb6428df177e8b597ff13e0216ebc2def1eed47177c18b7b8e5" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.264326 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cvrjq"] Mar 13 09:18:07 crc kubenswrapper[4930]: E0313 09:18:07.265087 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cddfdd5aa1f76cb6428df177e8b597ff13e0216ebc2def1eed47177c18b7b8e5\": container with ID starting with cddfdd5aa1f76cb6428df177e8b597ff13e0216ebc2def1eed47177c18b7b8e5 not found: ID does not exist" containerID="cddfdd5aa1f76cb6428df177e8b597ff13e0216ebc2def1eed47177c18b7b8e5" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.265114 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cddfdd5aa1f76cb6428df177e8b597ff13e0216ebc2def1eed47177c18b7b8e5"} err="failed to get container status \"cddfdd5aa1f76cb6428df177e8b597ff13e0216ebc2def1eed47177c18b7b8e5\": rpc error: code = NotFound desc = could not find container \"cddfdd5aa1f76cb6428df177e8b597ff13e0216ebc2def1eed47177c18b7b8e5\": container with ID starting with cddfdd5aa1f76cb6428df177e8b597ff13e0216ebc2def1eed47177c18b7b8e5 not found: ID does not exist" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.265128 4930 scope.go:117] "RemoveContainer" containerID="7ee210e26eaadde01e83f39b637dc5c600bd87d87402fcc3f3a29904b1b03c34" Mar 13 09:18:07 crc kubenswrapper[4930]: E0313 09:18:07.265975 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ee210e26eaadde01e83f39b637dc5c600bd87d87402fcc3f3a29904b1b03c34\": container with ID starting with 7ee210e26eaadde01e83f39b637dc5c600bd87d87402fcc3f3a29904b1b03c34 not found: ID does not exist" containerID="7ee210e26eaadde01e83f39b637dc5c600bd87d87402fcc3f3a29904b1b03c34" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.266000 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ee210e26eaadde01e83f39b637dc5c600bd87d87402fcc3f3a29904b1b03c34"} err="failed to get container status \"7ee210e26eaadde01e83f39b637dc5c600bd87d87402fcc3f3a29904b1b03c34\": rpc error: code = NotFound desc = could not find container \"7ee210e26eaadde01e83f39b637dc5c600bd87d87402fcc3f3a29904b1b03c34\": container with ID starting with 7ee210e26eaadde01e83f39b637dc5c600bd87d87402fcc3f3a29904b1b03c34 not found: ID does not exist" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.266015 4930 scope.go:117] "RemoveContainer" containerID="0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.271462 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cvrjq"] Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.281374 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tnq4n"] Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.289311 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tnq4n"] Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.294046 4930 scope.go:117] "RemoveContainer" containerID="cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.309598 4930 scope.go:117] "RemoveContainer" containerID="0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028" Mar 13 09:18:07 crc kubenswrapper[4930]: E0313 09:18:07.313586 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028\": container with ID starting with 0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028 not found: ID does not exist" containerID="0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.313631 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028"} err="failed to get container status \"0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028\": rpc error: code = NotFound desc = could not find container \"0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028\": container with ID starting with 0b7f264f8a64e1278c323f5fa35f6f8b9b5a6a42ea3926a84bbffe0d2925e028 not found: ID does not exist" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.313654 4930 scope.go:117] "RemoveContainer" containerID="cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe" Mar 13 09:18:07 crc kubenswrapper[4930]: E0313 09:18:07.314562 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe\": container with ID starting with cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe not found: ID does not exist" containerID="cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.314590 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe"} err="failed to get container status \"cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe\": rpc error: code = NotFound desc = could not find container \"cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe\": container with ID starting with cf07afb0cda64f4a9a998fc5371f89639877d673f2eeb9c6b9660490518ee7fe not found: ID does not exist" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.314605 4930 scope.go:117] "RemoveContainer" containerID="f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.328845 4930 scope.go:117] "RemoveContainer" containerID="3e4db90cb31b93c9af1338a9605a4d8528c3d6fd86c55011861513726d18c5cc" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.345593 4930 scope.go:117] "RemoveContainer" containerID="61171bdfcce9964a938bf78cd8d5640ee622e9e15edf8c00561b6c79ba36c665" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.361267 4930 scope.go:117] "RemoveContainer" containerID="f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e" Mar 13 09:18:07 crc kubenswrapper[4930]: E0313 09:18:07.361726 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e\": container with ID starting with f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e not found: ID does not exist" containerID="f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.361779 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e"} err="failed to get container status \"f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e\": rpc error: code = NotFound desc = could not find container \"f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e\": container with ID starting with f8d4109df76e46d8e9dbd0503659d5a486f535db6c90c237d42f552c90783c5e not found: ID does not exist" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.361801 4930 scope.go:117] "RemoveContainer" containerID="3e4db90cb31b93c9af1338a9605a4d8528c3d6fd86c55011861513726d18c5cc" Mar 13 09:18:07 crc kubenswrapper[4930]: E0313 09:18:07.362879 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e4db90cb31b93c9af1338a9605a4d8528c3d6fd86c55011861513726d18c5cc\": container with ID starting with 3e4db90cb31b93c9af1338a9605a4d8528c3d6fd86c55011861513726d18c5cc not found: ID does not exist" containerID="3e4db90cb31b93c9af1338a9605a4d8528c3d6fd86c55011861513726d18c5cc" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.362911 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e4db90cb31b93c9af1338a9605a4d8528c3d6fd86c55011861513726d18c5cc"} err="failed to get container status \"3e4db90cb31b93c9af1338a9605a4d8528c3d6fd86c55011861513726d18c5cc\": rpc error: code = NotFound desc = could not find container \"3e4db90cb31b93c9af1338a9605a4d8528c3d6fd86c55011861513726d18c5cc\": container with ID starting with 3e4db90cb31b93c9af1338a9605a4d8528c3d6fd86c55011861513726d18c5cc not found: ID does not exist" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.362924 4930 scope.go:117] "RemoveContainer" containerID="61171bdfcce9964a938bf78cd8d5640ee622e9e15edf8c00561b6c79ba36c665" Mar 13 09:18:07 crc kubenswrapper[4930]: E0313 09:18:07.363198 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61171bdfcce9964a938bf78cd8d5640ee622e9e15edf8c00561b6c79ba36c665\": container with ID starting with 61171bdfcce9964a938bf78cd8d5640ee622e9e15edf8c00561b6c79ba36c665 not found: ID does not exist" containerID="61171bdfcce9964a938bf78cd8d5640ee622e9e15edf8c00561b6c79ba36c665" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.363220 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61171bdfcce9964a938bf78cd8d5640ee622e9e15edf8c00561b6c79ba36c665"} err="failed to get container status \"61171bdfcce9964a938bf78cd8d5640ee622e9e15edf8c00561b6c79ba36c665\": rpc error: code = NotFound desc = could not find container \"61171bdfcce9964a938bf78cd8d5640ee622e9e15edf8c00561b6c79ba36c665\": container with ID starting with 61171bdfcce9964a938bf78cd8d5640ee622e9e15edf8c00561b6c79ba36c665 not found: ID does not exist" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.363232 4930 scope.go:117] "RemoveContainer" containerID="dcf841d84dca2fde34daf6d3dd2b33890ac1c4ae49782c2d6977ba7a83f68c5b" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.376839 4930 scope.go:117] "RemoveContainer" containerID="f0976d49697dba856874e4cf8f556c8b0961bc51aa9cb18c1ecda66745359dfe" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.389320 4930 scope.go:117] "RemoveContainer" containerID="5bccd725a88473e73b456ea970fe2fc850ccb07473af0c0d31d343287c681528" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.402228 4930 scope.go:117] "RemoveContainer" containerID="f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.413392 4930 scope.go:117] "RemoveContainer" containerID="f5f5b1abe368a8c35d77094d68c958aba293c94236448f088cc8e8498463ae52" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.424902 4930 scope.go:117] "RemoveContainer" containerID="472c78aa0cebb2e4a504a67acbafdf7e1bfb7cf49c0202dda94c0148623caec4" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.437897 4930 scope.go:117] "RemoveContainer" containerID="f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd" Mar 13 09:18:07 crc kubenswrapper[4930]: E0313 09:18:07.438155 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd\": container with ID starting with f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd not found: ID does not exist" containerID="f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.438184 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd"} err="failed to get container status \"f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd\": rpc error: code = NotFound desc = could not find container \"f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd\": container with ID starting with f1163249fb4ced13f8f773a638152a012307b1f3653f85ce6c5d9a5dabf48abd not found: ID does not exist" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.438206 4930 scope.go:117] "RemoveContainer" containerID="f5f5b1abe368a8c35d77094d68c958aba293c94236448f088cc8e8498463ae52" Mar 13 09:18:07 crc kubenswrapper[4930]: E0313 09:18:07.438392 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5f5b1abe368a8c35d77094d68c958aba293c94236448f088cc8e8498463ae52\": container with ID starting with f5f5b1abe368a8c35d77094d68c958aba293c94236448f088cc8e8498463ae52 not found: ID does not exist" containerID="f5f5b1abe368a8c35d77094d68c958aba293c94236448f088cc8e8498463ae52" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.438411 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5f5b1abe368a8c35d77094d68c958aba293c94236448f088cc8e8498463ae52"} err="failed to get container status \"f5f5b1abe368a8c35d77094d68c958aba293c94236448f088cc8e8498463ae52\": rpc error: code = NotFound desc = could not find container \"f5f5b1abe368a8c35d77094d68c958aba293c94236448f088cc8e8498463ae52\": container with ID starting with f5f5b1abe368a8c35d77094d68c958aba293c94236448f088cc8e8498463ae52 not found: ID does not exist" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.438425 4930 scope.go:117] "RemoveContainer" containerID="472c78aa0cebb2e4a504a67acbafdf7e1bfb7cf49c0202dda94c0148623caec4" Mar 13 09:18:07 crc kubenswrapper[4930]: E0313 09:18:07.438627 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"472c78aa0cebb2e4a504a67acbafdf7e1bfb7cf49c0202dda94c0148623caec4\": container with ID starting with 472c78aa0cebb2e4a504a67acbafdf7e1bfb7cf49c0202dda94c0148623caec4 not found: ID does not exist" containerID="472c78aa0cebb2e4a504a67acbafdf7e1bfb7cf49c0202dda94c0148623caec4" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.438645 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"472c78aa0cebb2e4a504a67acbafdf7e1bfb7cf49c0202dda94c0148623caec4"} err="failed to get container status \"472c78aa0cebb2e4a504a67acbafdf7e1bfb7cf49c0202dda94c0148623caec4\": rpc error: code = NotFound desc = could not find container \"472c78aa0cebb2e4a504a67acbafdf7e1bfb7cf49c0202dda94c0148623caec4\": container with ID starting with 472c78aa0cebb2e4a504a67acbafdf7e1bfb7cf49c0202dda94c0148623caec4 not found: ID does not exist" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.980819 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="085d31a1-d613-4e2f-a992-42bfc9b66413" path="/var/lib/kubelet/pods/085d31a1-d613-4e2f-a992-42bfc9b66413/volumes" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.982109 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" path="/var/lib/kubelet/pods/342671c3-b1f8-4c61-a0da-abe7dcdfc367/volumes" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.983516 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" path="/var/lib/kubelet/pods/5933708c-4e7f-4567-bd1e-df3dfe85fcd5/volumes" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.985839 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb104f34-0d73-40d9-9ba2-13401d61315d" path="/var/lib/kubelet/pods/bb104f34-0d73-40d9-9ba2-13401d61315d/volumes" Mar 13 09:18:07 crc kubenswrapper[4930]: I0313 09:18:07.986875 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf6e93be-455f-45e4-a51b-d851591c31ed" path="/var/lib/kubelet/pods/bf6e93be-455f-45e4-a51b-d851591c31ed/volumes" Mar 13 09:18:08 crc kubenswrapper[4930]: I0313 09:18:08.190288 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.022229 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kf9cp"] Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.022802 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerName="registry-server" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.022819 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerName="registry-server" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.022832 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb104f34-0d73-40d9-9ba2-13401d61315d" containerName="marketplace-operator" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.022841 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb104f34-0d73-40d9-9ba2-13401d61315d" containerName="marketplace-operator" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.022856 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6e93be-455f-45e4-a51b-d851591c31ed" containerName="extract-content" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.022863 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6e93be-455f-45e4-a51b-d851591c31ed" containerName="extract-content" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.022873 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" containerName="extract-utilities" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.022883 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" containerName="extract-utilities" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.022898 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb104f34-0d73-40d9-9ba2-13401d61315d" containerName="marketplace-operator" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.022905 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb104f34-0d73-40d9-9ba2-13401d61315d" containerName="marketplace-operator" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.022916 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerName="extract-content" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.022923 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerName="extract-content" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.022937 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerName="extract-utilities" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.022944 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerName="extract-utilities" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.022956 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6e93be-455f-45e4-a51b-d851591c31ed" containerName="registry-server" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.022963 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6e93be-455f-45e4-a51b-d851591c31ed" containerName="registry-server" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.022974 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" containerName="registry-server" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.022982 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" containerName="registry-server" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.022993 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" containerName="extract-content" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.023002 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" containerName="extract-content" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.023012 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085d31a1-d613-4e2f-a992-42bfc9b66413" containerName="extract-utilities" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.023019 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="085d31a1-d613-4e2f-a992-42bfc9b66413" containerName="extract-utilities" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.023028 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085d31a1-d613-4e2f-a992-42bfc9b66413" containerName="registry-server" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.023035 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="085d31a1-d613-4e2f-a992-42bfc9b66413" containerName="registry-server" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.023043 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf6e93be-455f-45e4-a51b-d851591c31ed" containerName="extract-utilities" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.023051 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf6e93be-455f-45e4-a51b-d851591c31ed" containerName="extract-utilities" Mar 13 09:18:09 crc kubenswrapper[4930]: E0313 09:18:09.023062 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="085d31a1-d613-4e2f-a992-42bfc9b66413" containerName="extract-content" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.023072 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="085d31a1-d613-4e2f-a992-42bfc9b66413" containerName="extract-content" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.023176 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb104f34-0d73-40d9-9ba2-13401d61315d" containerName="marketplace-operator" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.023190 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="5933708c-4e7f-4567-bd1e-df3dfe85fcd5" containerName="registry-server" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.023202 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb104f34-0d73-40d9-9ba2-13401d61315d" containerName="marketplace-operator" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.023213 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="085d31a1-d613-4e2f-a992-42bfc9b66413" containerName="registry-server" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.023224 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="342671c3-b1f8-4c61-a0da-abe7dcdfc367" containerName="registry-server" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.023234 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf6e93be-455f-45e4-a51b-d851591c31ed" containerName="registry-server" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.024124 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.026414 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.040624 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kf9cp"] Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.176604 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/240f1f92-8c63-4c78-b7a9-52e9133f4617-utilities\") pod \"redhat-marketplace-kf9cp\" (UID: \"240f1f92-8c63-4c78-b7a9-52e9133f4617\") " pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.176640 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/240f1f92-8c63-4c78-b7a9-52e9133f4617-catalog-content\") pod \"redhat-marketplace-kf9cp\" (UID: \"240f1f92-8c63-4c78-b7a9-52e9133f4617\") " pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.176659 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kxcq7\" (UniqueName: \"kubernetes.io/projected/240f1f92-8c63-4c78-b7a9-52e9133f4617-kube-api-access-kxcq7\") pod \"redhat-marketplace-kf9cp\" (UID: \"240f1f92-8c63-4c78-b7a9-52e9133f4617\") " pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.278047 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/240f1f92-8c63-4c78-b7a9-52e9133f4617-utilities\") pod \"redhat-marketplace-kf9cp\" (UID: \"240f1f92-8c63-4c78-b7a9-52e9133f4617\") " pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.278083 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/240f1f92-8c63-4c78-b7a9-52e9133f4617-catalog-content\") pod \"redhat-marketplace-kf9cp\" (UID: \"240f1f92-8c63-4c78-b7a9-52e9133f4617\") " pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.278106 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kxcq7\" (UniqueName: \"kubernetes.io/projected/240f1f92-8c63-4c78-b7a9-52e9133f4617-kube-api-access-kxcq7\") pod \"redhat-marketplace-kf9cp\" (UID: \"240f1f92-8c63-4c78-b7a9-52e9133f4617\") " pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.278668 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/240f1f92-8c63-4c78-b7a9-52e9133f4617-utilities\") pod \"redhat-marketplace-kf9cp\" (UID: \"240f1f92-8c63-4c78-b7a9-52e9133f4617\") " pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.278949 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/240f1f92-8c63-4c78-b7a9-52e9133f4617-catalog-content\") pod \"redhat-marketplace-kf9cp\" (UID: \"240f1f92-8c63-4c78-b7a9-52e9133f4617\") " pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.300976 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kxcq7\" (UniqueName: \"kubernetes.io/projected/240f1f92-8c63-4c78-b7a9-52e9133f4617-kube-api-access-kxcq7\") pod \"redhat-marketplace-kf9cp\" (UID: \"240f1f92-8c63-4c78-b7a9-52e9133f4617\") " pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.364634 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.585782 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kf9cp"] Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.618427 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-568x4"] Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.619397 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.623053 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.627173 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-568x4"] Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.685272 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dd00f07-858c-42be-a09e-746f709bbcdb-catalog-content\") pod \"redhat-operators-568x4\" (UID: \"4dd00f07-858c-42be-a09e-746f709bbcdb\") " pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.685341 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dd00f07-858c-42be-a09e-746f709bbcdb-utilities\") pod \"redhat-operators-568x4\" (UID: \"4dd00f07-858c-42be-a09e-746f709bbcdb\") " pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.685417 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2m2b\" (UniqueName: \"kubernetes.io/projected/4dd00f07-858c-42be-a09e-746f709bbcdb-kube-api-access-h2m2b\") pod \"redhat-operators-568x4\" (UID: \"4dd00f07-858c-42be-a09e-746f709bbcdb\") " pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.786972 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dd00f07-858c-42be-a09e-746f709bbcdb-catalog-content\") pod \"redhat-operators-568x4\" (UID: \"4dd00f07-858c-42be-a09e-746f709bbcdb\") " pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.787048 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dd00f07-858c-42be-a09e-746f709bbcdb-utilities\") pod \"redhat-operators-568x4\" (UID: \"4dd00f07-858c-42be-a09e-746f709bbcdb\") " pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.787140 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2m2b\" (UniqueName: \"kubernetes.io/projected/4dd00f07-858c-42be-a09e-746f709bbcdb-kube-api-access-h2m2b\") pod \"redhat-operators-568x4\" (UID: \"4dd00f07-858c-42be-a09e-746f709bbcdb\") " pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.788554 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4dd00f07-858c-42be-a09e-746f709bbcdb-utilities\") pod \"redhat-operators-568x4\" (UID: \"4dd00f07-858c-42be-a09e-746f709bbcdb\") " pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.788632 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4dd00f07-858c-42be-a09e-746f709bbcdb-catalog-content\") pod \"redhat-operators-568x4\" (UID: \"4dd00f07-858c-42be-a09e-746f709bbcdb\") " pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.805704 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2m2b\" (UniqueName: \"kubernetes.io/projected/4dd00f07-858c-42be-a09e-746f709bbcdb-kube-api-access-h2m2b\") pod \"redhat-operators-568x4\" (UID: \"4dd00f07-858c-42be-a09e-746f709bbcdb\") " pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:09 crc kubenswrapper[4930]: I0313 09:18:09.969060 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:10 crc kubenswrapper[4930]: I0313 09:18:10.195132 4930 generic.go:334] "Generic (PLEG): container finished" podID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerID="879669876b358fbbd78c70df6cfa466c3d2b15cb6da86e2d9d84677206a837b1" exitCode=0 Mar 13 09:18:10 crc kubenswrapper[4930]: I0313 09:18:10.195209 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kf9cp" event={"ID":"240f1f92-8c63-4c78-b7a9-52e9133f4617","Type":"ContainerDied","Data":"879669876b358fbbd78c70df6cfa466c3d2b15cb6da86e2d9d84677206a837b1"} Mar 13 09:18:10 crc kubenswrapper[4930]: I0313 09:18:10.195570 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kf9cp" event={"ID":"240f1f92-8c63-4c78-b7a9-52e9133f4617","Type":"ContainerStarted","Data":"ff51da5afe8b6e9a2a600a859fa6d27c4eb8160e77f5c8ea76ac29f345175381"} Mar 13 09:18:10 crc kubenswrapper[4930]: I0313 09:18:10.375454 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-568x4"] Mar 13 09:18:10 crc kubenswrapper[4930]: W0313 09:18:10.379166 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4dd00f07_858c_42be_a09e_746f709bbcdb.slice/crio-159326f701b321e38b663f06801b1f5accb00534366e17e6882437b78ae460f0 WatchSource:0}: Error finding container 159326f701b321e38b663f06801b1f5accb00534366e17e6882437b78ae460f0: Status 404 returned error can't find the container with id 159326f701b321e38b663f06801b1f5accb00534366e17e6882437b78ae460f0 Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.204686 4930 generic.go:334] "Generic (PLEG): container finished" podID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerID="e98ecf63fb77f0e72bc565312d0c0a08f2f3adf2731dd123cf18a1391bc3119e" exitCode=0 Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.204792 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kf9cp" event={"ID":"240f1f92-8c63-4c78-b7a9-52e9133f4617","Type":"ContainerDied","Data":"e98ecf63fb77f0e72bc565312d0c0a08f2f3adf2731dd123cf18a1391bc3119e"} Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.207029 4930 generic.go:334] "Generic (PLEG): container finished" podID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerID="e0198a1e4749403e1a32fa27bcb3c864e9ba0a78788408989fb6dcf88ee758b5" exitCode=0 Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.207083 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-568x4" event={"ID":"4dd00f07-858c-42be-a09e-746f709bbcdb","Type":"ContainerDied","Data":"e0198a1e4749403e1a32fa27bcb3c864e9ba0a78788408989fb6dcf88ee758b5"} Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.207122 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-568x4" event={"ID":"4dd00f07-858c-42be-a09e-746f709bbcdb","Type":"ContainerStarted","Data":"159326f701b321e38b663f06801b1f5accb00534366e17e6882437b78ae460f0"} Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.419638 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-z4n72"] Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.421339 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.423152 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.432691 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z4n72"] Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.618476 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/918d3d63-22c7-433d-86c8-22747f6d3e77-utilities\") pod \"community-operators-z4n72\" (UID: \"918d3d63-22c7-433d-86c8-22747f6d3e77\") " pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.618640 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/918d3d63-22c7-433d-86c8-22747f6d3e77-catalog-content\") pod \"community-operators-z4n72\" (UID: \"918d3d63-22c7-433d-86c8-22747f6d3e77\") " pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.618793 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bqtg\" (UniqueName: \"kubernetes.io/projected/918d3d63-22c7-433d-86c8-22747f6d3e77-kube-api-access-4bqtg\") pod \"community-operators-z4n72\" (UID: \"918d3d63-22c7-433d-86c8-22747f6d3e77\") " pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.725101 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bqtg\" (UniqueName: \"kubernetes.io/projected/918d3d63-22c7-433d-86c8-22747f6d3e77-kube-api-access-4bqtg\") pod \"community-operators-z4n72\" (UID: \"918d3d63-22c7-433d-86c8-22747f6d3e77\") " pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.725231 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/918d3d63-22c7-433d-86c8-22747f6d3e77-utilities\") pod \"community-operators-z4n72\" (UID: \"918d3d63-22c7-433d-86c8-22747f6d3e77\") " pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.725265 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/918d3d63-22c7-433d-86c8-22747f6d3e77-catalog-content\") pod \"community-operators-z4n72\" (UID: \"918d3d63-22c7-433d-86c8-22747f6d3e77\") " pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.725703 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/918d3d63-22c7-433d-86c8-22747f6d3e77-catalog-content\") pod \"community-operators-z4n72\" (UID: \"918d3d63-22c7-433d-86c8-22747f6d3e77\") " pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.725973 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/918d3d63-22c7-433d-86c8-22747f6d3e77-utilities\") pod \"community-operators-z4n72\" (UID: \"918d3d63-22c7-433d-86c8-22747f6d3e77\") " pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.750368 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bqtg\" (UniqueName: \"kubernetes.io/projected/918d3d63-22c7-433d-86c8-22747f6d3e77-kube-api-access-4bqtg\") pod \"community-operators-z4n72\" (UID: \"918d3d63-22c7-433d-86c8-22747f6d3e77\") " pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.765418 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" Mar 13 09:18:11 crc kubenswrapper[4930]: I0313 09:18:11.825080 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-shpj8"] Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.020685 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-62lp6"] Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.021836 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.024298 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.031364 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-62lp6"] Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.035614 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.129975 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/195e681c-d03e-42fb-8c90-b97ca4c380b2-utilities\") pod \"certified-operators-62lp6\" (UID: \"195e681c-d03e-42fb-8c90-b97ca4c380b2\") " pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.130321 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2fgf\" (UniqueName: \"kubernetes.io/projected/195e681c-d03e-42fb-8c90-b97ca4c380b2-kube-api-access-h2fgf\") pod \"certified-operators-62lp6\" (UID: \"195e681c-d03e-42fb-8c90-b97ca4c380b2\") " pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.130352 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/195e681c-d03e-42fb-8c90-b97ca4c380b2-catalog-content\") pod \"certified-operators-62lp6\" (UID: \"195e681c-d03e-42fb-8c90-b97ca4c380b2\") " pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.218334 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kf9cp" event={"ID":"240f1f92-8c63-4c78-b7a9-52e9133f4617","Type":"ContainerStarted","Data":"6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85"} Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.231051 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-568x4" event={"ID":"4dd00f07-858c-42be-a09e-746f709bbcdb","Type":"ContainerStarted","Data":"da13043f20d3a675d6b6496a4b1813100d6bd8d5c23719bf2b00e326b66ee2e1"} Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.231283 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/195e681c-d03e-42fb-8c90-b97ca4c380b2-utilities\") pod \"certified-operators-62lp6\" (UID: \"195e681c-d03e-42fb-8c90-b97ca4c380b2\") " pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.231348 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2fgf\" (UniqueName: \"kubernetes.io/projected/195e681c-d03e-42fb-8c90-b97ca4c380b2-kube-api-access-h2fgf\") pod \"certified-operators-62lp6\" (UID: \"195e681c-d03e-42fb-8c90-b97ca4c380b2\") " pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.231403 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/195e681c-d03e-42fb-8c90-b97ca4c380b2-catalog-content\") pod \"certified-operators-62lp6\" (UID: \"195e681c-d03e-42fb-8c90-b97ca4c380b2\") " pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.231943 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/195e681c-d03e-42fb-8c90-b97ca4c380b2-catalog-content\") pod \"certified-operators-62lp6\" (UID: \"195e681c-d03e-42fb-8c90-b97ca4c380b2\") " pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.232180 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/195e681c-d03e-42fb-8c90-b97ca4c380b2-utilities\") pod \"certified-operators-62lp6\" (UID: \"195e681c-d03e-42fb-8c90-b97ca4c380b2\") " pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.237467 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kf9cp" podStartSLOduration=1.833669403 podStartE2EDuration="3.237413101s" podCreationTimestamp="2026-03-13 09:18:09 +0000 UTC" firstStartedPulling="2026-03-13 09:18:10.197315119 +0000 UTC m=+330.947229806" lastFinishedPulling="2026-03-13 09:18:11.601058827 +0000 UTC m=+332.350973504" observedRunningTime="2026-03-13 09:18:12.231962977 +0000 UTC m=+332.981877664" watchObservedRunningTime="2026-03-13 09:18:12.237413101 +0000 UTC m=+332.987327778" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.255809 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2fgf\" (UniqueName: \"kubernetes.io/projected/195e681c-d03e-42fb-8c90-b97ca4c380b2-kube-api-access-h2fgf\") pod \"certified-operators-62lp6\" (UID: \"195e681c-d03e-42fb-8c90-b97ca4c380b2\") " pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.342119 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.456072 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-z4n72"] Mar 13 09:18:12 crc kubenswrapper[4930]: W0313 09:18:12.462735 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod918d3d63_22c7_433d_86c8_22747f6d3e77.slice/crio-7404131f69d6f13496d7a648edce854bd3b3b6057a056036b371524c1263dfbf WatchSource:0}: Error finding container 7404131f69d6f13496d7a648edce854bd3b3b6057a056036b371524c1263dfbf: Status 404 returned error can't find the container with id 7404131f69d6f13496d7a648edce854bd3b3b6057a056036b371524c1263dfbf Mar 13 09:18:12 crc kubenswrapper[4930]: I0313 09:18:12.548444 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-62lp6"] Mar 13 09:18:12 crc kubenswrapper[4930]: W0313 09:18:12.552590 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod195e681c_d03e_42fb_8c90_b97ca4c380b2.slice/crio-de8ee63bfee55f3f004ee9f4f5c8a22e26a0f58d6216d8b6d58eec4d1d8c9e91 WatchSource:0}: Error finding container de8ee63bfee55f3f004ee9f4f5c8a22e26a0f58d6216d8b6d58eec4d1d8c9e91: Status 404 returned error can't find the container with id de8ee63bfee55f3f004ee9f4f5c8a22e26a0f58d6216d8b6d58eec4d1d8c9e91 Mar 13 09:18:13 crc kubenswrapper[4930]: I0313 09:18:13.238732 4930 generic.go:334] "Generic (PLEG): container finished" podID="195e681c-d03e-42fb-8c90-b97ca4c380b2" containerID="672807ef8df3cce05223edfa1cf0b3de65042d4bf54f38610044352faf38b1ac" exitCode=0 Mar 13 09:18:13 crc kubenswrapper[4930]: I0313 09:18:13.238788 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62lp6" event={"ID":"195e681c-d03e-42fb-8c90-b97ca4c380b2","Type":"ContainerDied","Data":"672807ef8df3cce05223edfa1cf0b3de65042d4bf54f38610044352faf38b1ac"} Mar 13 09:18:13 crc kubenswrapper[4930]: I0313 09:18:13.238851 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62lp6" event={"ID":"195e681c-d03e-42fb-8c90-b97ca4c380b2","Type":"ContainerStarted","Data":"de8ee63bfee55f3f004ee9f4f5c8a22e26a0f58d6216d8b6d58eec4d1d8c9e91"} Mar 13 09:18:13 crc kubenswrapper[4930]: I0313 09:18:13.240478 4930 generic.go:334] "Generic (PLEG): container finished" podID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerID="da13043f20d3a675d6b6496a4b1813100d6bd8d5c23719bf2b00e326b66ee2e1" exitCode=0 Mar 13 09:18:13 crc kubenswrapper[4930]: I0313 09:18:13.240533 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-568x4" event={"ID":"4dd00f07-858c-42be-a09e-746f709bbcdb","Type":"ContainerDied","Data":"da13043f20d3a675d6b6496a4b1813100d6bd8d5c23719bf2b00e326b66ee2e1"} Mar 13 09:18:13 crc kubenswrapper[4930]: I0313 09:18:13.243305 4930 generic.go:334] "Generic (PLEG): container finished" podID="918d3d63-22c7-433d-86c8-22747f6d3e77" containerID="bacc91f65d692a20555d1d03af80b7ec87a0c091750d288483510ec0196a3660" exitCode=0 Mar 13 09:18:13 crc kubenswrapper[4930]: I0313 09:18:13.244624 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4n72" event={"ID":"918d3d63-22c7-433d-86c8-22747f6d3e77","Type":"ContainerDied","Data":"bacc91f65d692a20555d1d03af80b7ec87a0c091750d288483510ec0196a3660"} Mar 13 09:18:13 crc kubenswrapper[4930]: I0313 09:18:13.244660 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4n72" event={"ID":"918d3d63-22c7-433d-86c8-22747f6d3e77","Type":"ContainerStarted","Data":"7404131f69d6f13496d7a648edce854bd3b3b6057a056036b371524c1263dfbf"} Mar 13 09:18:14 crc kubenswrapper[4930]: I0313 09:18:14.264039 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62lp6" event={"ID":"195e681c-d03e-42fb-8c90-b97ca4c380b2","Type":"ContainerStarted","Data":"37c7eec8ff1306df653f7390596c310091d4e59a18db54f220a3ef68a9fbc7cb"} Mar 13 09:18:14 crc kubenswrapper[4930]: I0313 09:18:14.266564 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-568x4" event={"ID":"4dd00f07-858c-42be-a09e-746f709bbcdb","Type":"ContainerStarted","Data":"9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959"} Mar 13 09:18:14 crc kubenswrapper[4930]: I0313 09:18:14.269385 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4n72" event={"ID":"918d3d63-22c7-433d-86c8-22747f6d3e77","Type":"ContainerStarted","Data":"248a5fd5b70b5f6fcfaad1325a8b0cf6771095ef6f4792eae75699a9600dbf21"} Mar 13 09:18:14 crc kubenswrapper[4930]: I0313 09:18:14.331603 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-568x4" podStartSLOduration=2.844595485 podStartE2EDuration="5.331586341s" podCreationTimestamp="2026-03-13 09:18:09 +0000 UTC" firstStartedPulling="2026-03-13 09:18:11.210588838 +0000 UTC m=+331.960503555" lastFinishedPulling="2026-03-13 09:18:13.697579714 +0000 UTC m=+334.447494411" observedRunningTime="2026-03-13 09:18:14.327232994 +0000 UTC m=+335.077147661" watchObservedRunningTime="2026-03-13 09:18:14.331586341 +0000 UTC m=+335.081501028" Mar 13 09:18:15 crc kubenswrapper[4930]: I0313 09:18:15.276951 4930 generic.go:334] "Generic (PLEG): container finished" podID="918d3d63-22c7-433d-86c8-22747f6d3e77" containerID="248a5fd5b70b5f6fcfaad1325a8b0cf6771095ef6f4792eae75699a9600dbf21" exitCode=0 Mar 13 09:18:15 crc kubenswrapper[4930]: I0313 09:18:15.277029 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4n72" event={"ID":"918d3d63-22c7-433d-86c8-22747f6d3e77","Type":"ContainerDied","Data":"248a5fd5b70b5f6fcfaad1325a8b0cf6771095ef6f4792eae75699a9600dbf21"} Mar 13 09:18:15 crc kubenswrapper[4930]: I0313 09:18:15.278879 4930 generic.go:334] "Generic (PLEG): container finished" podID="195e681c-d03e-42fb-8c90-b97ca4c380b2" containerID="37c7eec8ff1306df653f7390596c310091d4e59a18db54f220a3ef68a9fbc7cb" exitCode=0 Mar 13 09:18:15 crc kubenswrapper[4930]: I0313 09:18:15.278948 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62lp6" event={"ID":"195e681c-d03e-42fb-8c90-b97ca4c380b2","Type":"ContainerDied","Data":"37c7eec8ff1306df653f7390596c310091d4e59a18db54f220a3ef68a9fbc7cb"} Mar 13 09:18:16 crc kubenswrapper[4930]: I0313 09:18:16.286767 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4n72" event={"ID":"918d3d63-22c7-433d-86c8-22747f6d3e77","Type":"ContainerStarted","Data":"afac97426e2bd5db823c7655dde77c9560e76c4e36127f86764ec3dc68f6a106"} Mar 13 09:18:16 crc kubenswrapper[4930]: I0313 09:18:16.289804 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-62lp6" event={"ID":"195e681c-d03e-42fb-8c90-b97ca4c380b2","Type":"ContainerStarted","Data":"bf9844e518f9b0c78e33e56b35b82e623581b642c50db0a85b9cf51c47a3d960"} Mar 13 09:18:16 crc kubenswrapper[4930]: I0313 09:18:16.324746 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-z4n72" podStartSLOduration=2.869571374 podStartE2EDuration="5.324731034s" podCreationTimestamp="2026-03-13 09:18:11 +0000 UTC" firstStartedPulling="2026-03-13 09:18:13.245294307 +0000 UTC m=+333.995209014" lastFinishedPulling="2026-03-13 09:18:15.700453997 +0000 UTC m=+336.450368674" observedRunningTime="2026-03-13 09:18:16.323154025 +0000 UTC m=+337.073068702" watchObservedRunningTime="2026-03-13 09:18:16.324731034 +0000 UTC m=+337.074645711" Mar 13 09:18:16 crc kubenswrapper[4930]: I0313 09:18:16.350976 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-62lp6" podStartSLOduration=1.923745042 podStartE2EDuration="4.350948211s" podCreationTimestamp="2026-03-13 09:18:12 +0000 UTC" firstStartedPulling="2026-03-13 09:18:13.240876338 +0000 UTC m=+333.990791015" lastFinishedPulling="2026-03-13 09:18:15.668079467 +0000 UTC m=+336.417994184" observedRunningTime="2026-03-13 09:18:16.346055551 +0000 UTC m=+337.095970238" watchObservedRunningTime="2026-03-13 09:18:16.350948211 +0000 UTC m=+337.100862928" Mar 13 09:18:19 crc kubenswrapper[4930]: I0313 09:18:19.365670 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:19 crc kubenswrapper[4930]: I0313 09:18:19.365757 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:19 crc kubenswrapper[4930]: I0313 09:18:19.427259 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:19 crc kubenswrapper[4930]: I0313 09:18:19.970960 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:19 crc kubenswrapper[4930]: I0313 09:18:19.971032 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:20 crc kubenswrapper[4930]: I0313 09:18:20.348850 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 09:18:21 crc kubenswrapper[4930]: I0313 09:18:21.031092 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output=< Mar 13 09:18:21 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:18:21 crc kubenswrapper[4930]: > Mar 13 09:18:22 crc kubenswrapper[4930]: I0313 09:18:22.035866 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:22 crc kubenswrapper[4930]: I0313 09:18:22.035930 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:22 crc kubenswrapper[4930]: I0313 09:18:22.100951 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:22 crc kubenswrapper[4930]: I0313 09:18:22.342815 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:22 crc kubenswrapper[4930]: I0313 09:18:22.342873 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:22 crc kubenswrapper[4930]: I0313 09:18:22.382753 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:18:22 crc kubenswrapper[4930]: I0313 09:18:22.400553 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:23 crc kubenswrapper[4930]: I0313 09:18:23.406986 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-62lp6" Mar 13 09:18:30 crc kubenswrapper[4930]: I0313 09:18:30.034749 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:30 crc kubenswrapper[4930]: I0313 09:18:30.091686 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-568x4" Mar 13 09:18:36 crc kubenswrapper[4930]: I0313 09:18:36.865277 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" podUID="22f82073-0b04-4c7a-ab16-e88878214407" containerName="registry" containerID="cri-o://f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd" gracePeriod=30 Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.196553 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.307312 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/22f82073-0b04-4c7a-ab16-e88878214407-installation-pull-secrets\") pod \"22f82073-0b04-4c7a-ab16-e88878214407\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.307383 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k472w\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-kube-api-access-k472w\") pod \"22f82073-0b04-4c7a-ab16-e88878214407\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.307445 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-registry-tls\") pod \"22f82073-0b04-4c7a-ab16-e88878214407\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.307493 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-bound-sa-token\") pod \"22f82073-0b04-4c7a-ab16-e88878214407\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.307573 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/22f82073-0b04-4c7a-ab16-e88878214407-registry-certificates\") pod \"22f82073-0b04-4c7a-ab16-e88878214407\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.307774 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"22f82073-0b04-4c7a-ab16-e88878214407\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.307825 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22f82073-0b04-4c7a-ab16-e88878214407-trusted-ca\") pod \"22f82073-0b04-4c7a-ab16-e88878214407\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.307858 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/22f82073-0b04-4c7a-ab16-e88878214407-ca-trust-extracted\") pod \"22f82073-0b04-4c7a-ab16-e88878214407\" (UID: \"22f82073-0b04-4c7a-ab16-e88878214407\") " Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.311236 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22f82073-0b04-4c7a-ab16-e88878214407-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "22f82073-0b04-4c7a-ab16-e88878214407" (UID: "22f82073-0b04-4c7a-ab16-e88878214407"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.311245 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22f82073-0b04-4c7a-ab16-e88878214407-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "22f82073-0b04-4c7a-ab16-e88878214407" (UID: "22f82073-0b04-4c7a-ab16-e88878214407"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.316418 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "22f82073-0b04-4c7a-ab16-e88878214407" (UID: "22f82073-0b04-4c7a-ab16-e88878214407"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.316509 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22f82073-0b04-4c7a-ab16-e88878214407-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "22f82073-0b04-4c7a-ab16-e88878214407" (UID: "22f82073-0b04-4c7a-ab16-e88878214407"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.325696 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "22f82073-0b04-4c7a-ab16-e88878214407" (UID: "22f82073-0b04-4c7a-ab16-e88878214407"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.325826 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "22f82073-0b04-4c7a-ab16-e88878214407" (UID: "22f82073-0b04-4c7a-ab16-e88878214407"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.325927 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-kube-api-access-k472w" (OuterVolumeSpecName: "kube-api-access-k472w") pod "22f82073-0b04-4c7a-ab16-e88878214407" (UID: "22f82073-0b04-4c7a-ab16-e88878214407"). InnerVolumeSpecName "kube-api-access-k472w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.329756 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22f82073-0b04-4c7a-ab16-e88878214407-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "22f82073-0b04-4c7a-ab16-e88878214407" (UID: "22f82073-0b04-4c7a-ab16-e88878214407"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.409626 4930 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/22f82073-0b04-4c7a-ab16-e88878214407-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.409870 4930 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/22f82073-0b04-4c7a-ab16-e88878214407-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.409933 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k472w\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-kube-api-access-k472w\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.409994 4930 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.410060 4930 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/22f82073-0b04-4c7a-ab16-e88878214407-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.410125 4930 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/22f82073-0b04-4c7a-ab16-e88878214407-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.410185 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/22f82073-0b04-4c7a-ab16-e88878214407-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.429737 4930 generic.go:334] "Generic (PLEG): container finished" podID="22f82073-0b04-4c7a-ab16-e88878214407" containerID="f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd" exitCode=0 Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.429772 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" event={"ID":"22f82073-0b04-4c7a-ab16-e88878214407","Type":"ContainerDied","Data":"f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd"} Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.429798 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" event={"ID":"22f82073-0b04-4c7a-ab16-e88878214407","Type":"ContainerDied","Data":"b684c160c05b89dd54570071a7a23c9b05bc0b5f6ef44fe653231a7015e9f6f5"} Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.429818 4930 scope.go:117] "RemoveContainer" containerID="f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.429815 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-shpj8" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.445263 4930 scope.go:117] "RemoveContainer" containerID="f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd" Mar 13 09:18:37 crc kubenswrapper[4930]: E0313 09:18:37.446468 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd\": container with ID starting with f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd not found: ID does not exist" containerID="f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.446494 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd"} err="failed to get container status \"f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd\": rpc error: code = NotFound desc = could not find container \"f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd\": container with ID starting with f0a2a275a8492d84b5c67cbfc916e21d01a22b5caf66d9b019934c6d07ade9fd not found: ID does not exist" Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.459159 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-shpj8"] Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.463649 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-shpj8"] Mar 13 09:18:37 crc kubenswrapper[4930]: I0313 09:18:37.977479 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22f82073-0b04-4c7a-ab16-e88878214407" path="/var/lib/kubelet/pods/22f82073-0b04-4c7a-ab16-e88878214407/volumes" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.295765 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz"] Mar 13 09:18:59 crc kubenswrapper[4930]: E0313 09:18:59.296722 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22f82073-0b04-4c7a-ab16-e88878214407" containerName="registry" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.296745 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="22f82073-0b04-4c7a-ab16-e88878214407" containerName="registry" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.296928 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="22f82073-0b04-4c7a-ab16-e88878214407" containerName="registry" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.297584 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.299844 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-tls" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.301074 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-root-ca.crt" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.301260 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"telemetry-config" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.301696 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"cluster-monitoring-operator-dockercfg-wwt9l" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.302063 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"openshift-service-ca.crt" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.315396 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz"] Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.408367 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/7b0c297e-f66a-4036-8dc0-02e6fdd5f851-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-26dgz\" (UID: \"7b0c297e-f66a-4036-8dc0-02e6fdd5f851\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.409287 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v944f\" (UniqueName: \"kubernetes.io/projected/7b0c297e-f66a-4036-8dc0-02e6fdd5f851-kube-api-access-v944f\") pod \"cluster-monitoring-operator-6d5b84845-26dgz\" (UID: \"7b0c297e-f66a-4036-8dc0-02e6fdd5f851\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.409639 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/7b0c297e-f66a-4036-8dc0-02e6fdd5f851-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-26dgz\" (UID: \"7b0c297e-f66a-4036-8dc0-02e6fdd5f851\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.518355 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/7b0c297e-f66a-4036-8dc0-02e6fdd5f851-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-26dgz\" (UID: \"7b0c297e-f66a-4036-8dc0-02e6fdd5f851\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.518757 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v944f\" (UniqueName: \"kubernetes.io/projected/7b0c297e-f66a-4036-8dc0-02e6fdd5f851-kube-api-access-v944f\") pod \"cluster-monitoring-operator-6d5b84845-26dgz\" (UID: \"7b0c297e-f66a-4036-8dc0-02e6fdd5f851\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.518898 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/7b0c297e-f66a-4036-8dc0-02e6fdd5f851-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-26dgz\" (UID: \"7b0c297e-f66a-4036-8dc0-02e6fdd5f851\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.524065 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-config\" (UniqueName: \"kubernetes.io/configmap/7b0c297e-f66a-4036-8dc0-02e6fdd5f851-telemetry-config\") pod \"cluster-monitoring-operator-6d5b84845-26dgz\" (UID: \"7b0c297e-f66a-4036-8dc0-02e6fdd5f851\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.537468 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cluster-monitoring-operator-tls\" (UniqueName: \"kubernetes.io/secret/7b0c297e-f66a-4036-8dc0-02e6fdd5f851-cluster-monitoring-operator-tls\") pod \"cluster-monitoring-operator-6d5b84845-26dgz\" (UID: \"7b0c297e-f66a-4036-8dc0-02e6fdd5f851\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.544986 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v944f\" (UniqueName: \"kubernetes.io/projected/7b0c297e-f66a-4036-8dc0-02e6fdd5f851-kube-api-access-v944f\") pod \"cluster-monitoring-operator-6d5b84845-26dgz\" (UID: \"7b0c297e-f66a-4036-8dc0-02e6fdd5f851\") " pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.617993 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" Mar 13 09:18:59 crc kubenswrapper[4930]: I0313 09:18:59.844517 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz"] Mar 13 09:19:00 crc kubenswrapper[4930]: I0313 09:19:00.574041 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" event={"ID":"7b0c297e-f66a-4036-8dc0-02e6fdd5f851","Type":"ContainerStarted","Data":"32ae225119d75b4c7842a0dc87df9bf90c3b7e35ecd767a82a5e7d24f2ef5484"} Mar 13 09:19:02 crc kubenswrapper[4930]: I0313 09:19:02.346453 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k"] Mar 13 09:19:02 crc kubenswrapper[4930]: I0313 09:19:02.347598 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" Mar 13 09:19:02 crc kubenswrapper[4930]: I0313 09:19:02.349897 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-dockercfg-rgjln" Mar 13 09:19:02 crc kubenswrapper[4930]: I0313 09:19:02.350005 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-admission-webhook-tls" Mar 13 09:19:02 crc kubenswrapper[4930]: I0313 09:19:02.359990 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k"] Mar 13 09:19:02 crc kubenswrapper[4930]: I0313 09:19:02.457343 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/8f964d06-4f78-4e56-ac62-fc19a2020489-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-zfz2k\" (UID: \"8f964d06-4f78-4e56-ac62-fc19a2020489\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" Mar 13 09:19:02 crc kubenswrapper[4930]: I0313 09:19:02.558939 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/8f964d06-4f78-4e56-ac62-fc19a2020489-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-zfz2k\" (UID: \"8f964d06-4f78-4e56-ac62-fc19a2020489\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" Mar 13 09:19:02 crc kubenswrapper[4930]: I0313 09:19:02.564563 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-certificates\" (UniqueName: \"kubernetes.io/secret/8f964d06-4f78-4e56-ac62-fc19a2020489-tls-certificates\") pod \"prometheus-operator-admission-webhook-f54c54754-zfz2k\" (UID: \"8f964d06-4f78-4e56-ac62-fc19a2020489\") " pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" Mar 13 09:19:02 crc kubenswrapper[4930]: I0313 09:19:02.585826 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" event={"ID":"7b0c297e-f66a-4036-8dc0-02e6fdd5f851","Type":"ContainerStarted","Data":"393d6b37d0a61738ca157ee33e0ec6cda26610af241c731b5e6e7e3a556c3337"} Mar 13 09:19:02 crc kubenswrapper[4930]: I0313 09:19:02.667634 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" Mar 13 09:19:02 crc kubenswrapper[4930]: I0313 09:19:02.845642 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/cluster-monitoring-operator-6d5b84845-26dgz" podStartSLOduration=1.98831943 podStartE2EDuration="3.845599226s" podCreationTimestamp="2026-03-13 09:18:59 +0000 UTC" firstStartedPulling="2026-03-13 09:18:59.850649278 +0000 UTC m=+380.600563955" lastFinishedPulling="2026-03-13 09:19:01.707929074 +0000 UTC m=+382.457843751" observedRunningTime="2026-03-13 09:19:02.604359355 +0000 UTC m=+383.354274042" watchObservedRunningTime="2026-03-13 09:19:02.845599226 +0000 UTC m=+383.595513903" Mar 13 09:19:02 crc kubenswrapper[4930]: I0313 09:19:02.847534 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k"] Mar 13 09:19:02 crc kubenswrapper[4930]: W0313 09:19:02.852770 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8f964d06_4f78_4e56_ac62_fc19a2020489.slice/crio-b67fceacc84ab295503ed2ee0837e6b10cc9005833c080e6151007ca54f332e3 WatchSource:0}: Error finding container b67fceacc84ab295503ed2ee0837e6b10cc9005833c080e6151007ca54f332e3: Status 404 returned error can't find the container with id b67fceacc84ab295503ed2ee0837e6b10cc9005833c080e6151007ca54f332e3 Mar 13 09:19:03 crc kubenswrapper[4930]: I0313 09:19:03.594642 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" event={"ID":"8f964d06-4f78-4e56-ac62-fc19a2020489","Type":"ContainerStarted","Data":"b67fceacc84ab295503ed2ee0837e6b10cc9005833c080e6151007ca54f332e3"} Mar 13 09:19:04 crc kubenswrapper[4930]: I0313 09:19:04.603717 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" event={"ID":"8f964d06-4f78-4e56-ac62-fc19a2020489","Type":"ContainerStarted","Data":"4f86d8a5b91c950152ea5feebeab50790467d609ad31c9a05c0b392b87c01c02"} Mar 13 09:19:04 crc kubenswrapper[4930]: I0313 09:19:04.604069 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" Mar 13 09:19:04 crc kubenswrapper[4930]: I0313 09:19:04.611920 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" Mar 13 09:19:04 crc kubenswrapper[4930]: I0313 09:19:04.625994 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" podStartSLOduration=1.143650121 podStartE2EDuration="2.625975761s" podCreationTimestamp="2026-03-13 09:19:02 +0000 UTC" firstStartedPulling="2026-03-13 09:19:02.854944967 +0000 UTC m=+383.604859644" lastFinishedPulling="2026-03-13 09:19:04.337270607 +0000 UTC m=+385.087185284" observedRunningTime="2026-03-13 09:19:04.624147606 +0000 UTC m=+385.374062323" watchObservedRunningTime="2026-03-13 09:19:04.625975761 +0000 UTC m=+385.375890448" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.420305 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-wq9rb"] Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.421125 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.424236 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-kube-rbac-proxy-config" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.424629 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-client-ca" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.425143 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-tls" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.425859 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-operator-dockercfg-tjv72" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.437297 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-wq9rb"] Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.601671 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9cmxx\" (UniqueName: \"kubernetes.io/projected/eab5c5a2-7c78-4e65-857a-7a5da73f24a3-kube-api-access-9cmxx\") pod \"prometheus-operator-db54df47d-wq9rb\" (UID: \"eab5c5a2-7c78-4e65-857a-7a5da73f24a3\") " pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.601721 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/eab5c5a2-7c78-4e65-857a-7a5da73f24a3-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-wq9rb\" (UID: \"eab5c5a2-7c78-4e65-857a-7a5da73f24a3\") " pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.601743 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/eab5c5a2-7c78-4e65-857a-7a5da73f24a3-metrics-client-ca\") pod \"prometheus-operator-db54df47d-wq9rb\" (UID: \"eab5c5a2-7c78-4e65-857a-7a5da73f24a3\") " pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.601946 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/eab5c5a2-7c78-4e65-857a-7a5da73f24a3-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-wq9rb\" (UID: \"eab5c5a2-7c78-4e65-857a-7a5da73f24a3\") " pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.703118 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9cmxx\" (UniqueName: \"kubernetes.io/projected/eab5c5a2-7c78-4e65-857a-7a5da73f24a3-kube-api-access-9cmxx\") pod \"prometheus-operator-db54df47d-wq9rb\" (UID: \"eab5c5a2-7c78-4e65-857a-7a5da73f24a3\") " pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.703876 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/eab5c5a2-7c78-4e65-857a-7a5da73f24a3-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-wq9rb\" (UID: \"eab5c5a2-7c78-4e65-857a-7a5da73f24a3\") " pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.703940 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/eab5c5a2-7c78-4e65-857a-7a5da73f24a3-metrics-client-ca\") pod \"prometheus-operator-db54df47d-wq9rb\" (UID: \"eab5c5a2-7c78-4e65-857a-7a5da73f24a3\") " pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.704065 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/eab5c5a2-7c78-4e65-857a-7a5da73f24a3-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-wq9rb\" (UID: \"eab5c5a2-7c78-4e65-857a-7a5da73f24a3\") " pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.705728 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/eab5c5a2-7c78-4e65-857a-7a5da73f24a3-metrics-client-ca\") pod \"prometheus-operator-db54df47d-wq9rb\" (UID: \"eab5c5a2-7c78-4e65-857a-7a5da73f24a3\") " pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.712096 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/eab5c5a2-7c78-4e65-857a-7a5da73f24a3-prometheus-operator-kube-rbac-proxy-config\") pod \"prometheus-operator-db54df47d-wq9rb\" (UID: \"eab5c5a2-7c78-4e65-857a-7a5da73f24a3\") " pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.712140 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-operator-tls\" (UniqueName: \"kubernetes.io/secret/eab5c5a2-7c78-4e65-857a-7a5da73f24a3-prometheus-operator-tls\") pod \"prometheus-operator-db54df47d-wq9rb\" (UID: \"eab5c5a2-7c78-4e65-857a-7a5da73f24a3\") " pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.718932 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9cmxx\" (UniqueName: \"kubernetes.io/projected/eab5c5a2-7c78-4e65-857a-7a5da73f24a3-kube-api-access-9cmxx\") pod \"prometheus-operator-db54df47d-wq9rb\" (UID: \"eab5c5a2-7c78-4e65-857a-7a5da73f24a3\") " pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:05 crc kubenswrapper[4930]: I0313 09:19:05.736237 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" Mar 13 09:19:06 crc kubenswrapper[4930]: I0313 09:19:06.163577 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-operator-db54df47d-wq9rb"] Mar 13 09:19:06 crc kubenswrapper[4930]: W0313 09:19:06.170744 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeab5c5a2_7c78_4e65_857a_7a5da73f24a3.slice/crio-0e657a6c2487cdb547620c1ac5ebbd6e4bb18b304f868b0270d8b55f9159a44c WatchSource:0}: Error finding container 0e657a6c2487cdb547620c1ac5ebbd6e4bb18b304f868b0270d8b55f9159a44c: Status 404 returned error can't find the container with id 0e657a6c2487cdb547620c1ac5ebbd6e4bb18b304f868b0270d8b55f9159a44c Mar 13 09:19:06 crc kubenswrapper[4930]: I0313 09:19:06.623149 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" event={"ID":"eab5c5a2-7c78-4e65-857a-7a5da73f24a3","Type":"ContainerStarted","Data":"0e657a6c2487cdb547620c1ac5ebbd6e4bb18b304f868b0270d8b55f9159a44c"} Mar 13 09:19:08 crc kubenswrapper[4930]: I0313 09:19:08.635573 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" event={"ID":"eab5c5a2-7c78-4e65-857a-7a5da73f24a3","Type":"ContainerStarted","Data":"ac67a498752165009ef4bcad531b9e82b0d3b40db3fe7547b3120eb1da8ba0fa"} Mar 13 09:19:08 crc kubenswrapper[4930]: I0313 09:19:08.635892 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" event={"ID":"eab5c5a2-7c78-4e65-857a-7a5da73f24a3","Type":"ContainerStarted","Data":"9b89b4e006467cdc674db819e05924636755f3c68d8b60be7275f00481ed55b3"} Mar 13 09:19:08 crc kubenswrapper[4930]: I0313 09:19:08.654135 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-operator-db54df47d-wq9rb" podStartSLOduration=1.651331771 podStartE2EDuration="3.654113913s" podCreationTimestamp="2026-03-13 09:19:05 +0000 UTC" firstStartedPulling="2026-03-13 09:19:06.17260391 +0000 UTC m=+386.922518587" lastFinishedPulling="2026-03-13 09:19:08.175386052 +0000 UTC m=+388.925300729" observedRunningTime="2026-03-13 09:19:08.649847807 +0000 UTC m=+389.399762484" watchObservedRunningTime="2026-03-13 09:19:08.654113913 +0000 UTC m=+389.404028590" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.796114 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh"] Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.797674 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.799285 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-tls" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.799295 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-kube-rbac-proxy-config" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.803381 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"openshift-state-metrics-dockercfg-gfx62" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.817192 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh"] Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.824684 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh"] Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.826005 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.827968 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-tls" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.828082 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-dockercfg-ztzvh" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.828257 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kube-state-metrics-custom-resource-state-configmap" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.828781 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-state-metrics-kube-rbac-proxy-config" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.850469 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh"] Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.882130 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/node-exporter-b4h2q"] Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.891644 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.899793 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-tls" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.899930 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-kube-rbac-proxy-config" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.900734 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"node-exporter-dockercfg-z4fm5" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.975153 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.975202 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.975439 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkpt7\" (UniqueName: \"kubernetes.io/projected/5f98c586-64b7-4173-82f4-b86b6ab66f04-kube-api-access-wkpt7\") pod \"openshift-state-metrics-566fddb674-9h5bh\" (UID: \"5f98c586-64b7-4173-82f4-b86b6ab66f04\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.975477 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/5f98c586-64b7-4173-82f4-b86b6ab66f04-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-9h5bh\" (UID: \"5f98c586-64b7-4173-82f4-b86b6ab66f04\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.975500 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5f98c586-64b7-4173-82f4-b86b6ab66f04-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-9h5bh\" (UID: \"5f98c586-64b7-4173-82f4-b86b6ab66f04\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.975520 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g55l9\" (UniqueName: \"kubernetes.io/projected/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-api-access-g55l9\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.975620 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/eea04747-6fe4-4ecd-b310-a6bf2e57952f-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.975696 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/eea04747-6fe4-4ecd-b310-a6bf2e57952f-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.975762 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/5f98c586-64b7-4173-82f4-b86b6ab66f04-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-9h5bh\" (UID: \"5f98c586-64b7-4173-82f4-b86b6ab66f04\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:10 crc kubenswrapper[4930]: I0313 09:19:10.975879 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077259 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5f98c586-64b7-4173-82f4-b86b6ab66f04-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-9h5bh\" (UID: \"5f98c586-64b7-4173-82f4-b86b6ab66f04\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077317 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a743b119-af9e-46cd-8a3f-56b888e4b90e-sys\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077349 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g55l9\" (UniqueName: \"kubernetes.io/projected/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-api-access-g55l9\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077413 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/eea04747-6fe4-4ecd-b310-a6bf2e57952f-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077460 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/eea04747-6fe4-4ecd-b310-a6bf2e57952f-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077488 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-tls\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077520 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vc22\" (UniqueName: \"kubernetes.io/projected/a743b119-af9e-46cd-8a3f-56b888e4b90e-kube-api-access-8vc22\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077554 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/5f98c586-64b7-4173-82f4-b86b6ab66f04-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-9h5bh\" (UID: \"5f98c586-64b7-4173-82f4-b86b6ab66f04\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077599 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/a743b119-af9e-46cd-8a3f-56b888e4b90e-root\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077639 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-textfile\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077710 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077742 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-wtmp\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077836 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077860 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.077883 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkpt7\" (UniqueName: \"kubernetes.io/projected/5f98c586-64b7-4173-82f4-b86b6ab66f04-kube-api-access-wkpt7\") pod \"openshift-state-metrics-566fddb674-9h5bh\" (UID: \"5f98c586-64b7-4173-82f4-b86b6ab66f04\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.078024 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"volume-directive-shadow\" (UniqueName: \"kubernetes.io/empty-dir/eea04747-6fe4-4ecd-b310-a6bf2e57952f-volume-directive-shadow\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: E0313 09:19:11.078321 4930 secret.go:188] Couldn't get secret openshift-monitoring/kube-state-metrics-tls: secret "kube-state-metrics-tls" not found Mar 13 09:19:11 crc kubenswrapper[4930]: E0313 09:19:11.078371 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-state-metrics-tls podName:eea04747-6fe4-4ecd-b310-a6bf2e57952f nodeName:}" failed. No retries permitted until 2026-03-13 09:19:11.578355219 +0000 UTC m=+392.328269896 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-state-metrics-tls" (UniqueName: "kubernetes.io/secret/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-state-metrics-tls") pod "kube-state-metrics-777cb5bd5d-f2bmh" (UID: "eea04747-6fe4-4ecd-b310-a6bf2e57952f") : secret "kube-state-metrics-tls" not found Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.078389 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.078450 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5f98c586-64b7-4173-82f4-b86b6ab66f04-metrics-client-ca\") pod \"openshift-state-metrics-566fddb674-9h5bh\" (UID: \"5f98c586-64b7-4173-82f4-b86b6ab66f04\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.078467 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a743b119-af9e-46cd-8a3f-56b888e4b90e-metrics-client-ca\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.078496 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/eea04747-6fe4-4ecd-b310-a6bf2e57952f-metrics-client-ca\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.078508 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/5f98c586-64b7-4173-82f4-b86b6ab66f04-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-9h5bh\" (UID: \"5f98c586-64b7-4173-82f4-b86b6ab66f04\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.079187 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-custom-resource-state-configmap\" (UniqueName: \"kubernetes.io/configmap/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-state-metrics-custom-resource-state-configmap\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.083810 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/5f98c586-64b7-4173-82f4-b86b6ab66f04-openshift-state-metrics-kube-rbac-proxy-config\") pod \"openshift-state-metrics-566fddb674-9h5bh\" (UID: \"5f98c586-64b7-4173-82f4-b86b6ab66f04\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.083809 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-state-metrics-kube-rbac-proxy-config\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.084329 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/5f98c586-64b7-4173-82f4-b86b6ab66f04-openshift-state-metrics-tls\") pod \"openshift-state-metrics-566fddb674-9h5bh\" (UID: \"5f98c586-64b7-4173-82f4-b86b6ab66f04\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.094808 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g55l9\" (UniqueName: \"kubernetes.io/projected/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-api-access-g55l9\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.096861 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkpt7\" (UniqueName: \"kubernetes.io/projected/5f98c586-64b7-4173-82f4-b86b6ab66f04-kube-api-access-wkpt7\") pod \"openshift-state-metrics-566fddb674-9h5bh\" (UID: \"5f98c586-64b7-4173-82f4-b86b6ab66f04\") " pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.113368 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.180243 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-tls\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.180301 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vc22\" (UniqueName: \"kubernetes.io/projected/a743b119-af9e-46cd-8a3f-56b888e4b90e-kube-api-access-8vc22\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.180352 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"root\" (UniqueName: \"kubernetes.io/host-path/a743b119-af9e-46cd-8a3f-56b888e4b90e-root\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.180387 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-textfile\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.180411 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-wtmp\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: E0313 09:19:11.180448 4930 secret.go:188] Couldn't get secret openshift-monitoring/node-exporter-tls: secret "node-exporter-tls" not found Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.180470 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.180496 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a743b119-af9e-46cd-8a3f-56b888e4b90e-metrics-client-ca\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: E0313 09:19:11.180529 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-tls podName:a743b119-af9e-46cd-8a3f-56b888e4b90e nodeName:}" failed. No retries permitted until 2026-03-13 09:19:11.680505383 +0000 UTC m=+392.430420060 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-exporter-tls" (UniqueName: "kubernetes.io/secret/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-tls") pod "node-exporter-b4h2q" (UID: "a743b119-af9e-46cd-8a3f-56b888e4b90e") : secret "node-exporter-tls" not found Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.180563 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a743b119-af9e-46cd-8a3f-56b888e4b90e-sys\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.180628 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a743b119-af9e-46cd-8a3f-56b888e4b90e-sys\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.180808 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"root\" (UniqueName: \"kubernetes.io/host-path/a743b119-af9e-46cd-8a3f-56b888e4b90e-root\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.180914 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-wtmp\" (UniqueName: \"kubernetes.io/host-path/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-wtmp\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.181296 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-textfile\" (UniqueName: \"kubernetes.io/empty-dir/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-textfile\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.181367 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/a743b119-af9e-46cd-8a3f-56b888e4b90e-metrics-client-ca\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.189091 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-kube-rbac-proxy-config\" (UniqueName: \"kubernetes.io/secret/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-kube-rbac-proxy-config\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.194670 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vc22\" (UniqueName: \"kubernetes.io/projected/a743b119-af9e-46cd-8a3f-56b888e4b90e-kube-api-access-8vc22\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.364826 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh"] Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.585813 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.593008 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls\" (UniqueName: \"kubernetes.io/secret/eea04747-6fe4-4ecd-b310-a6bf2e57952f-kube-state-metrics-tls\") pod \"kube-state-metrics-777cb5bd5d-f2bmh\" (UID: \"eea04747-6fe4-4ecd-b310-a6bf2e57952f\") " pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.652503 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" event={"ID":"5f98c586-64b7-4173-82f4-b86b6ab66f04","Type":"ContainerStarted","Data":"35f6b2915ae70c14009b36db26ab13f2ec62a4a1eb7529df3975399952b2f2dd"} Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.652852 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" event={"ID":"5f98c586-64b7-4173-82f4-b86b6ab66f04","Type":"ContainerStarted","Data":"ad559f132cfeecfbca5ea889a3fd0d08ba3317b324953c33a0f80e251e7ce149"} Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.652867 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" event={"ID":"5f98c586-64b7-4173-82f4-b86b6ab66f04","Type":"ContainerStarted","Data":"7de0b4de635b62b7b405ed7abd469ba4f79d4045d41cff9166f073f913cd77ff"} Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.686243 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-tls\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.689529 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-exporter-tls\" (UniqueName: \"kubernetes.io/secret/a743b119-af9e-46cd-8a3f-56b888e4b90e-node-exporter-tls\") pod \"node-exporter-b4h2q\" (UID: \"a743b119-af9e-46cd-8a3f-56b888e4b90e\") " pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.740464 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.819925 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/node-exporter-b4h2q" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.903789 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.908407 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.910452 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-metric" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.912468 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-web-config" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.912682 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-dockercfg-fkchb" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.912887 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.913776 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-generated" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.913932 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-main-tls-assets-0" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.914035 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy-web" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.919582 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"alertmanager-kube-rbac-proxy" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.921833 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"alertmanager-trusted-ca-bundle" Mar 13 09:19:11 crc kubenswrapper[4930]: I0313 09:19:11.954378 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.008064 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.008113 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.008230 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-config-out\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.008309 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.008474 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-web-config\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.008541 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.008600 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.008727 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.008756 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-config-volume\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.008823 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.008859 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5762\" (UniqueName: \"kubernetes.io/projected/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-kube-api-access-p5762\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.008911 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-tls-assets\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.109745 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.110301 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-web-config\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.110332 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.110356 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.110382 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.110399 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-config-volume\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.110424 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.110462 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5762\" (UniqueName: \"kubernetes.io/projected/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-kube-api-access-p5762\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.110488 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-tls-assets\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.110520 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.110542 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.110568 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-config-out\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.111340 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-main-db\" (UniqueName: \"kubernetes.io/empty-dir/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-alertmanager-main-db\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.111904 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-metrics-client-ca\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.112019 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"alertmanager-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-alertmanager-trusted-ca-bundle\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.115327 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-config-volume\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.115378 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-config-out\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.115374 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-secret-alertmanager-kube-rbac-proxy\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.115532 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-secret-alertmanager-kube-rbac-proxy-web\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.116022 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-main-tls\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-secret-alertmanager-main-tls\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.117450 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-web-config\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.118282 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-alertmanager-kube-rbac-proxy-metric\" (UniqueName: \"kubernetes.io/secret/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-secret-alertmanager-kube-rbac-proxy-metric\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.118640 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-tls-assets\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.130161 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5762\" (UniqueName: \"kubernetes.io/projected/4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749-kube-api-access-p5762\") pod \"alertmanager-main-0\" (UID: \"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749\") " pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.193401 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh"] Mar 13 09:19:12 crc kubenswrapper[4930]: W0313 09:19:12.196737 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeea04747_6fe4_4ecd_b310_a6bf2e57952f.slice/crio-5cfcce50304bf92ecad52caac69ac9195bab819f186cc36b5d77538c2cf3442f WatchSource:0}: Error finding container 5cfcce50304bf92ecad52caac69ac9195bab819f186cc36b5d77538c2cf3442f: Status 404 returned error can't find the container with id 5cfcce50304bf92ecad52caac69ac9195bab819f186cc36b5d77538c2cf3442f Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.226654 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/alertmanager-main-0" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.308458 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.308515 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.639362 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/alertmanager-main-0"] Mar 13 09:19:12 crc kubenswrapper[4930]: W0313 09:19:12.645959 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4fd42ac8_5bf8_4112_9d7a_0b67d9c8e749.slice/crio-a38031062dfb686252e69f81449ec98cfb49d90c193e1beb9eb64ed648ecb13a WatchSource:0}: Error finding container a38031062dfb686252e69f81449ec98cfb49d90c193e1beb9eb64ed648ecb13a: Status 404 returned error can't find the container with id a38031062dfb686252e69f81449ec98cfb49d90c193e1beb9eb64ed648ecb13a Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.660661 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" event={"ID":"eea04747-6fe4-4ecd-b310-a6bf2e57952f","Type":"ContainerStarted","Data":"5cfcce50304bf92ecad52caac69ac9195bab819f186cc36b5d77538c2cf3442f"} Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.663770 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-b4h2q" event={"ID":"a743b119-af9e-46cd-8a3f-56b888e4b90e","Type":"ContainerStarted","Data":"d53ee8a9f9f32af06d9fc490e02e306813800989f49279d32c8229de03631244"} Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.665623 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749","Type":"ContainerStarted","Data":"a38031062dfb686252e69f81449ec98cfb49d90c193e1beb9eb64ed648ecb13a"} Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.782555 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/thanos-querier-7c466dcfcb-k7psd"] Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.784378 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.788991 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.789339 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-tls" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.789554 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-dockercfg-9ws2h" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.789764 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-grpc-tls-3paj008moht2o" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.789906 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-rules" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.790093 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-metrics" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.790299 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"thanos-querier-kube-rbac-proxy-web" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.802023 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-7c466dcfcb-k7psd"] Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.919502 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-metrics-client-ca\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.919594 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-grpc-tls\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.919642 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.919699 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq9m6\" (UniqueName: \"kubernetes.io/projected/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-kube-api-access-cq9m6\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.920114 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.920224 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.920383 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-tls\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:12 crc kubenswrapper[4930]: I0313 09:19:12.920606 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.021763 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-metrics-client-ca\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.021848 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-grpc-tls\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.021880 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.021946 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq9m6\" (UniqueName: \"kubernetes.io/projected/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-kube-api-access-cq9m6\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.022008 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.022076 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.022157 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-tls\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.022201 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.024566 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-metrics-client-ca\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.029248 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-metrics\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-kube-rbac-proxy-metrics\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.030680 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-tls\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-tls\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.030826 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-grpc-tls\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.031364 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-rules\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-kube-rbac-proxy-rules\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.032334 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-kube-rbac-proxy-web\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.036128 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-thanos-querier-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-secret-thanos-querier-kube-rbac-proxy\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.045090 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq9m6\" (UniqueName: \"kubernetes.io/projected/951e59b6-d39a-4ec2-beda-4b6b8b3e20c8-kube-api-access-cq9m6\") pod \"thanos-querier-7c466dcfcb-k7psd\" (UID: \"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8\") " pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.103092 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:13 crc kubenswrapper[4930]: I0313 09:19:13.527488 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/thanos-querier-7c466dcfcb-k7psd"] Mar 13 09:19:13 crc kubenswrapper[4930]: W0313 09:19:13.919724 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod951e59b6_d39a_4ec2_beda_4b6b8b3e20c8.slice/crio-8bbe81c5e47c4a0e495232172c63a797c78691d5abcf62023529523a434e06e4 WatchSource:0}: Error finding container 8bbe81c5e47c4a0e495232172c63a797c78691d5abcf62023529523a434e06e4: Status 404 returned error can't find the container with id 8bbe81c5e47c4a0e495232172c63a797c78691d5abcf62023529523a434e06e4 Mar 13 09:19:14 crc kubenswrapper[4930]: I0313 09:19:14.678601 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" event={"ID":"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8","Type":"ContainerStarted","Data":"8bbe81c5e47c4a0e495232172c63a797c78691d5abcf62023529523a434e06e4"} Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.587446 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-568fc7fbc4-zps72"] Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.588685 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.605161 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-568fc7fbc4-zps72"] Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.668818 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-console-config\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.669088 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gn8w4\" (UniqueName: \"kubernetes.io/projected/80440d2f-0dac-42ca-a98a-855c116a80fd-kube-api-access-gn8w4\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.669105 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/80440d2f-0dac-42ca-a98a-855c116a80fd-console-oauth-config\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.669128 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-oauth-serving-cert\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.669161 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-service-ca\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.669294 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/80440d2f-0dac-42ca-a98a-855c116a80fd-console-serving-cert\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.669331 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-trusted-ca-bundle\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.684916 4930 generic.go:334] "Generic (PLEG): container finished" podID="4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749" containerID="a2e766278e8b50de0918b0dc58dbfba0150c46ad30efbcd1aa1796865b84009c" exitCode=0 Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.684974 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749","Type":"ContainerDied","Data":"a2e766278e8b50de0918b0dc58dbfba0150c46ad30efbcd1aa1796865b84009c"} Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.686616 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" event={"ID":"eea04747-6fe4-4ecd-b310-a6bf2e57952f","Type":"ContainerStarted","Data":"f39e51144466d2f3808c398201e20429a64da2b476a6effb8b326e0bad3509f1"} Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.686638 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" event={"ID":"eea04747-6fe4-4ecd-b310-a6bf2e57952f","Type":"ContainerStarted","Data":"32d78d97ea5f0999d8a01a9cc0c54bea05181636dfea63c110cd725a1795e78d"} Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.686650 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" event={"ID":"eea04747-6fe4-4ecd-b310-a6bf2e57952f","Type":"ContainerStarted","Data":"c21b89df988bf784d0fd142908e5adb44648c45c9c02a947204bbbccf86dd48c"} Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.687975 4930 generic.go:334] "Generic (PLEG): container finished" podID="a743b119-af9e-46cd-8a3f-56b888e4b90e" containerID="64e648c01f0b8b291ca8d8bd1582709f89d910f877a33409aac0b32d92996ea5" exitCode=0 Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.688033 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-b4h2q" event={"ID":"a743b119-af9e-46cd-8a3f-56b888e4b90e","Type":"ContainerDied","Data":"64e648c01f0b8b291ca8d8bd1582709f89d910f877a33409aac0b32d92996ea5"} Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.692300 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" event={"ID":"5f98c586-64b7-4173-82f4-b86b6ab66f04","Type":"ContainerStarted","Data":"b1f77a2897ec68a76e220d90e076f3f5d76aaf009d2c7a04d239a68f1d6728cf"} Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.755377 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/kube-state-metrics-777cb5bd5d-f2bmh" podStartSLOduration=3.440198193 podStartE2EDuration="5.755359048s" podCreationTimestamp="2026-03-13 09:19:10 +0000 UTC" firstStartedPulling="2026-03-13 09:19:12.198720931 +0000 UTC m=+392.948635608" lastFinishedPulling="2026-03-13 09:19:14.513881796 +0000 UTC m=+395.263796463" observedRunningTime="2026-03-13 09:19:15.754010064 +0000 UTC m=+396.503924741" watchObservedRunningTime="2026-03-13 09:19:15.755359048 +0000 UTC m=+396.505273725" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.770177 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/80440d2f-0dac-42ca-a98a-855c116a80fd-console-serving-cert\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.770222 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-trusted-ca-bundle\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.770256 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-console-config\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.771216 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-console-config\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.771282 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gn8w4\" (UniqueName: \"kubernetes.io/projected/80440d2f-0dac-42ca-a98a-855c116a80fd-kube-api-access-gn8w4\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.771312 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/80440d2f-0dac-42ca-a98a-855c116a80fd-console-oauth-config\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.771350 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-trusted-ca-bundle\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.771385 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-oauth-serving-cert\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.771531 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-service-ca\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.772588 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-service-ca\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.773117 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-oauth-serving-cert\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.777326 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/80440d2f-0dac-42ca-a98a-855c116a80fd-console-serving-cert\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.785171 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/openshift-state-metrics-566fddb674-9h5bh" podStartSLOduration=2.874446377 podStartE2EDuration="5.785151984s" podCreationTimestamp="2026-03-13 09:19:10 +0000 UTC" firstStartedPulling="2026-03-13 09:19:11.597547979 +0000 UTC m=+392.347462656" lastFinishedPulling="2026-03-13 09:19:14.508253596 +0000 UTC m=+395.258168263" observedRunningTime="2026-03-13 09:19:15.783973425 +0000 UTC m=+396.533888112" watchObservedRunningTime="2026-03-13 09:19:15.785151984 +0000 UTC m=+396.535066661" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.786074 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/80440d2f-0dac-42ca-a98a-855c116a80fd-console-oauth-config\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.796946 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gn8w4\" (UniqueName: \"kubernetes.io/projected/80440d2f-0dac-42ca-a98a-855c116a80fd-kube-api-access-gn8w4\") pod \"console-568fc7fbc4-zps72\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:15 crc kubenswrapper[4930]: I0313 09:19:15.902071 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.195535 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/metrics-server-647f59f4f4-vnqd8"] Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.196523 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.198202 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"kubelet-serving-ca-bundle" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.198649 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-client-certs" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.198694 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-tls" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.198822 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-dockercfg-kqqjq" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.198914 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"metrics-server-audit-profiles" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.198929 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"metrics-server-77ueeundo00ih" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.211164 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-647f59f4f4-vnqd8"] Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.266006 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-568fc7fbc4-zps72"] Mar 13 09:19:16 crc kubenswrapper[4930]: W0313 09:19:16.275648 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80440d2f_0dac_42ca_a98a_855c116a80fd.slice/crio-5488f0b2e11e7ffe4d94877af41242fcee30f3a547619d65c73c0f7b93275e00 WatchSource:0}: Error finding container 5488f0b2e11e7ffe4d94877af41242fcee30f3a547619d65c73c0f7b93275e00: Status 404 returned error can't find the container with id 5488f0b2e11e7ffe4d94877af41242fcee30f3a547619d65c73c0f7b93275e00 Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.380606 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-client-ca-bundle\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.380925 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-secret-metrics-client-certs\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.380958 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-audit-log\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.380983 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-metrics-server-audit-profiles\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.381009 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxpjz\" (UniqueName: \"kubernetes.io/projected/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-kube-api-access-cxpjz\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.381029 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.381060 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-secret-metrics-server-tls\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.482373 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-secret-metrics-client-certs\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.482465 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-audit-log\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.482498 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-metrics-server-audit-profiles\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.482533 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxpjz\" (UniqueName: \"kubernetes.io/projected/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-kube-api-access-cxpjz\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.482569 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-secret-metrics-server-tls\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.482594 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.482658 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-client-ca-bundle\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.483138 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-log\" (UniqueName: \"kubernetes.io/empty-dir/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-audit-log\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.483836 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-configmap-kubelet-serving-ca-bundle\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.484053 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-server-audit-profiles\" (UniqueName: \"kubernetes.io/configmap/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-metrics-server-audit-profiles\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.489114 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-secret-metrics-client-certs\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.490465 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-server-tls\" (UniqueName: \"kubernetes.io/secret/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-secret-metrics-server-tls\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.490894 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-client-ca-bundle\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.502216 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxpjz\" (UniqueName: \"kubernetes.io/projected/e4d5bf77-4258-4108-ad1b-f51fa45a3bab-kube-api-access-cxpjz\") pod \"metrics-server-647f59f4f4-vnqd8\" (UID: \"e4d5bf77-4258-4108-ad1b-f51fa45a3bab\") " pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.517510 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.582220 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f"] Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.585105 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.594029 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"monitoring-plugin-cert" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.594585 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"default-dockercfg-6tstp" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.599137 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f"] Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.685030 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/e52e16c5-4758-47c8-b6a3-0c2998455bd3-monitoring-plugin-cert\") pod \"monitoring-plugin-7f784d5d74-hnt2f\" (UID: \"e52e16c5-4758-47c8-b6a3-0c2998455bd3\") " pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.699508 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-568fc7fbc4-zps72" event={"ID":"80440d2f-0dac-42ca-a98a-855c116a80fd","Type":"ContainerStarted","Data":"44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe"} Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.699552 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-568fc7fbc4-zps72" event={"ID":"80440d2f-0dac-42ca-a98a-855c116a80fd","Type":"ContainerStarted","Data":"5488f0b2e11e7ffe4d94877af41242fcee30f3a547619d65c73c0f7b93275e00"} Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.703483 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-b4h2q" event={"ID":"a743b119-af9e-46cd-8a3f-56b888e4b90e","Type":"ContainerStarted","Data":"f4fd62c5e6cd51458b892c27bf64c8da8740e23caa331c8f38a544235856ac68"} Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.703511 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/node-exporter-b4h2q" event={"ID":"a743b119-af9e-46cd-8a3f-56b888e4b90e","Type":"ContainerStarted","Data":"3035365d5348c468521a89cbd7598a118b7d0410cae610c49d52ed11986af4cf"} Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.720147 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-568fc7fbc4-zps72" podStartSLOduration=1.7201288479999999 podStartE2EDuration="1.720128848s" podCreationTimestamp="2026-03-13 09:19:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:19:16.716907358 +0000 UTC m=+397.466822035" watchObservedRunningTime="2026-03-13 09:19:16.720128848 +0000 UTC m=+397.470043525" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.788165 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/e52e16c5-4758-47c8-b6a3-0c2998455bd3-monitoring-plugin-cert\") pod \"monitoring-plugin-7f784d5d74-hnt2f\" (UID: \"e52e16c5-4758-47c8-b6a3-0c2998455bd3\") " pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.792803 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/node-exporter-b4h2q" podStartSLOduration=4.15849297 podStartE2EDuration="6.792782214s" podCreationTimestamp="2026-03-13 09:19:10 +0000 UTC" firstStartedPulling="2026-03-13 09:19:11.873616074 +0000 UTC m=+392.623530751" lastFinishedPulling="2026-03-13 09:19:14.507905318 +0000 UTC m=+395.257819995" observedRunningTime="2026-03-13 09:19:16.749942525 +0000 UTC m=+397.499857202" watchObservedRunningTime="2026-03-13 09:19:16.792782214 +0000 UTC m=+397.542696891" Mar 13 09:19:16 crc kubenswrapper[4930]: W0313 09:19:16.805822 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4d5bf77_4258_4108_ad1b_f51fa45a3bab.slice/crio-34ad92720f21fa53f03999fa3c75f37b91a272da373060a1188690308c63f1c4 WatchSource:0}: Error finding container 34ad92720f21fa53f03999fa3c75f37b91a272da373060a1188690308c63f1c4: Status 404 returned error can't find the container with id 34ad92720f21fa53f03999fa3c75f37b91a272da373060a1188690308c63f1c4 Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.808641 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/metrics-server-647f59f4f4-vnqd8"] Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.809223 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"monitoring-plugin-cert\" (UniqueName: \"kubernetes.io/secret/e52e16c5-4758-47c8-b6a3-0c2998455bd3-monitoring-plugin-cert\") pod \"monitoring-plugin-7f784d5d74-hnt2f\" (UID: \"e52e16c5-4758-47c8-b6a3-0c2998455bd3\") " pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" Mar 13 09:19:16 crc kubenswrapper[4930]: I0313 09:19:16.952151 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.093415 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.095351 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.099201 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-dockercfg-pr57l" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.099630 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls-assets-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.099651 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-tls" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.099739 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"serving-certs-ca-bundle" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.099753 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-sidecar-tls" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.099836 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-thanos-prometheus-http-client-file" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.100419 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-web-config" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.100524 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-kube-rbac-proxy-web" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.100601 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s-grpc-tls-806qr2lh9iu9i" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.100982 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"prometheus-k8s" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.101094 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-monitoring"/"kube-rbac-proxy" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.105327 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-k8s-rulefiles-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.113362 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-monitoring"/"prometheus-trusted-ca-bundle" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.116541 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197094 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-config\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197141 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkwkd\" (UniqueName: \"kubernetes.io/projected/5f969789-8156-442b-b177-685669da5416-kube-api-access-kkwkd\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197172 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197196 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197226 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197251 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197273 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/5f969789-8156-442b-b177-685669da5416-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197296 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197333 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197359 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197373 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-web-config\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197391 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197409 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5f969789-8156-442b-b177-685669da5416-config-out\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197426 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197457 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197475 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5f969789-8156-442b-b177-685669da5416-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197490 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.197508 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.298943 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/5f969789-8156-442b-b177-685669da5416-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299022 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299071 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299100 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299117 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-web-config\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299138 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299157 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5f969789-8156-442b-b177-685669da5416-config-out\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299174 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5f969789-8156-442b-b177-685669da5416-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299192 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299212 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299236 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299259 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299278 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-config\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299296 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkwkd\" (UniqueName: \"kubernetes.io/projected/5f969789-8156-442b-b177-685669da5416-kube-api-access-kkwkd\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299315 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299331 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299347 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299363 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.299566 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-db\" (UniqueName: \"kubernetes.io/empty-dir/5f969789-8156-442b-b177-685669da5416-prometheus-k8s-db\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.305294 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-web-config\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.306687 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-kube-rbac-proxy\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-kube-rbac-proxy\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.307058 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-prometheus-trusted-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.307197 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-kubelet-serving-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-configmap-kubelet-serving-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.307384 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-thanos-sidecar-tls\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-prometheus-k8s-thanos-sidecar-tls\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.309686 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-config\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.310594 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/5f969789-8156-442b-b177-685669da5416-tls-assets\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.311352 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-serving-certs-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-configmap-serving-certs-ca-bundle\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.311714 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/5f969789-8156-442b-b177-685669da5416-config-out\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.312050 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-tls\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-prometheus-k8s-tls\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.316667 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-thanos-prometheus-http-client-file\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.317160 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-metrics-client-certs\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-metrics-client-certs\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.319111 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-grpc-tls\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-grpc-tls\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.319172 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"configmap-metrics-client-ca\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-configmap-metrics-client-ca\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.322266 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-prometheus-k8s-kube-rbac-proxy-web\" (UniqueName: \"kubernetes.io/secret/5f969789-8156-442b-b177-685669da5416-secret-prometheus-k8s-kube-rbac-proxy-web\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.328459 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-k8s-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/5f969789-8156-442b-b177-685669da5416-prometheus-k8s-rulefiles-0\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.336056 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkwkd\" (UniqueName: \"kubernetes.io/projected/5f969789-8156-442b-b177-685669da5416-kube-api-access-kkwkd\") pod \"prometheus-k8s-0\" (UID: \"5f969789-8156-442b-b177-685669da5416\") " pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.401209 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f"] Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.412472 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.636429 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-monitoring/prometheus-k8s-0"] Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.715142 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" event={"ID":"e52e16c5-4758-47c8-b6a3-0c2998455bd3","Type":"ContainerStarted","Data":"a14931dc206becafecda8fc8d0d14be815b5396a46c481f5c2ba752f6f79cb24"} Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.716937 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5f969789-8156-442b-b177-685669da5416","Type":"ContainerStarted","Data":"54898eb0344abb8ec73c77cb8bc5661a3df95221fc2107a2c4427c8dca576b59"} Mar 13 09:19:17 crc kubenswrapper[4930]: I0313 09:19:17.719298 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" event={"ID":"e4d5bf77-4258-4108-ad1b-f51fa45a3bab","Type":"ContainerStarted","Data":"34ad92720f21fa53f03999fa3c75f37b91a272da373060a1188690308c63f1c4"} Mar 13 09:19:18 crc kubenswrapper[4930]: I0313 09:19:18.725853 4930 generic.go:334] "Generic (PLEG): container finished" podID="5f969789-8156-442b-b177-685669da5416" containerID="c8f3a727710d9ea8b132dbdee8a1bdc26fb408d414fc4a667d8eb009d0354c01" exitCode=0 Mar 13 09:19:18 crc kubenswrapper[4930]: I0313 09:19:18.725908 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5f969789-8156-442b-b177-685669da5416","Type":"ContainerDied","Data":"c8f3a727710d9ea8b132dbdee8a1bdc26fb408d414fc4a667d8eb009d0354c01"} Mar 13 09:19:19 crc kubenswrapper[4930]: I0313 09:19:19.733638 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" event={"ID":"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8","Type":"ContainerStarted","Data":"1ad191e66a9b2530b0f0137e365478643b523ca4443460bf597df96dec3b3139"} Mar 13 09:19:19 crc kubenswrapper[4930]: I0313 09:19:19.734032 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" event={"ID":"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8","Type":"ContainerStarted","Data":"600a249418071c916b1151a202432aa9de42ad892d5aa455fe0b3a4ce05b9fe2"} Mar 13 09:19:19 crc kubenswrapper[4930]: I0313 09:19:19.735592 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" event={"ID":"e4d5bf77-4258-4108-ad1b-f51fa45a3bab","Type":"ContainerStarted","Data":"5620f3a4a3a7c68ab03d8e44a6f4f2ed95ed87191aa5c9da5fa88541afd99535"} Mar 13 09:19:19 crc kubenswrapper[4930]: I0313 09:19:19.739279 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749","Type":"ContainerStarted","Data":"f55e8108b583b79d09d54d66d22f82874fad29272469eeb6c43f83646f672b6c"} Mar 13 09:19:19 crc kubenswrapper[4930]: I0313 09:19:19.739321 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749","Type":"ContainerStarted","Data":"cc6b20f9d77ae5f3b16ec602d3b69b3cdea4d780c89b5b43dba7d4244242d76a"} Mar 13 09:19:19 crc kubenswrapper[4930]: I0313 09:19:19.757463 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" podStartSLOduration=1.195172842 podStartE2EDuration="3.757445527s" podCreationTimestamp="2026-03-13 09:19:16 +0000 UTC" firstStartedPulling="2026-03-13 09:19:16.813226279 +0000 UTC m=+397.563140956" lastFinishedPulling="2026-03-13 09:19:19.375498924 +0000 UTC m=+400.125413641" observedRunningTime="2026-03-13 09:19:19.750764612 +0000 UTC m=+400.500679299" watchObservedRunningTime="2026-03-13 09:19:19.757445527 +0000 UTC m=+400.507360204" Mar 13 09:19:20 crc kubenswrapper[4930]: I0313 09:19:20.747511 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749","Type":"ContainerStarted","Data":"6f95a39b863602f1e6f98283ae3a47b1cbcc68752995c08d6cfc1331709ee41c"} Mar 13 09:19:20 crc kubenswrapper[4930]: I0313 09:19:20.747760 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749","Type":"ContainerStarted","Data":"fe709c856c04f9f0494aad0b0c52450e02b7ac84391b4ce500ecc0e4a7456039"} Mar 13 09:19:20 crc kubenswrapper[4930]: I0313 09:19:20.752163 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" event={"ID":"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8","Type":"ContainerStarted","Data":"678dfedd652f209af9cac5749ad0e5ddbb27e6c8415fcc8c49b3e2f434eace69"} Mar 13 09:19:21 crc kubenswrapper[4930]: I0313 09:19:21.761975 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749","Type":"ContainerStarted","Data":"8f33461b694969e20f5fcaf812864fee37b7560b9b725ab1626f3b51d347c192"} Mar 13 09:19:21 crc kubenswrapper[4930]: I0313 09:19:21.764284 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" event={"ID":"e52e16c5-4758-47c8-b6a3-0c2998455bd3","Type":"ContainerStarted","Data":"a8d23ec81608701392ef0b13c56bad95bdc10fc75d6dbe151d2c5b3376978384"} Mar 13 09:19:21 crc kubenswrapper[4930]: I0313 09:19:21.764536 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" Mar 13 09:19:21 crc kubenswrapper[4930]: I0313 09:19:21.768747 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" Mar 13 09:19:21 crc kubenswrapper[4930]: I0313 09:19:21.776630 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" podStartSLOduration=2.504952542 podStartE2EDuration="5.776615624s" podCreationTimestamp="2026-03-13 09:19:16 +0000 UTC" firstStartedPulling="2026-03-13 09:19:17.413330145 +0000 UTC m=+398.163244822" lastFinishedPulling="2026-03-13 09:19:20.684993227 +0000 UTC m=+401.434907904" observedRunningTime="2026-03-13 09:19:21.775464936 +0000 UTC m=+402.525379613" watchObservedRunningTime="2026-03-13 09:19:21.776615624 +0000 UTC m=+402.526530301" Mar 13 09:19:23 crc kubenswrapper[4930]: I0313 09:19:23.779453 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5f969789-8156-442b-b177-685669da5416","Type":"ContainerStarted","Data":"ff2b2e4b07d1743606f85dd1e55b0ebd6149138c72ecefaae553c266e27b575e"} Mar 13 09:19:23 crc kubenswrapper[4930]: I0313 09:19:23.779784 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5f969789-8156-442b-b177-685669da5416","Type":"ContainerStarted","Data":"1687564f2e703b32a639e036942f5c7161b2659f59f0f165489b992552ee549a"} Mar 13 09:19:24 crc kubenswrapper[4930]: I0313 09:19:24.789068 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5f969789-8156-442b-b177-685669da5416","Type":"ContainerStarted","Data":"ee136f33d853b9bcecc58b1645cd928e55b7d3111031725d1d399fc9e07e5014"} Mar 13 09:19:24 crc kubenswrapper[4930]: I0313 09:19:24.789413 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5f969789-8156-442b-b177-685669da5416","Type":"ContainerStarted","Data":"d00215db72d890c1deca1d943f7e4929a9522505ddcf4fde52464eba81a89792"} Mar 13 09:19:25 crc kubenswrapper[4930]: I0313 09:19:25.802886 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/alertmanager-main-0" event={"ID":"4fd42ac8-5bf8-4112-9d7a-0b67d9c8e749","Type":"ContainerStarted","Data":"7d9d3497cb5e0fecc7f987585a9ff8aaa76287d633c3054441b9671304a6ff17"} Mar 13 09:19:25 crc kubenswrapper[4930]: I0313 09:19:25.812141 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" event={"ID":"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8","Type":"ContainerStarted","Data":"b1a8549fd71a61968c2d3f8fed991cc0cfe7f49b9908945be9337eacf4dca7c7"} Mar 13 09:19:25 crc kubenswrapper[4930]: I0313 09:19:25.812199 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" event={"ID":"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8","Type":"ContainerStarted","Data":"dbd103d997becbdf73691a3ef22b231070a3785b863d372989c6273fdfa5e2eb"} Mar 13 09:19:25 crc kubenswrapper[4930]: I0313 09:19:25.818783 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5f969789-8156-442b-b177-685669da5416","Type":"ContainerStarted","Data":"7ed32d515fc3b0658601856bb8954418d9764a880608ae74205ca0c06a5b8010"} Mar 13 09:19:25 crc kubenswrapper[4930]: I0313 09:19:25.818839 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5f969789-8156-442b-b177-685669da5416","Type":"ContainerStarted","Data":"aae428da127548b80600cf4eab956b5368fee04aceac0399059bdcdf8cfc63c5"} Mar 13 09:19:25 crc kubenswrapper[4930]: I0313 09:19:25.857317 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/alertmanager-main-0" podStartSLOduration=2.3991921290000002 podStartE2EDuration="14.857292936s" podCreationTimestamp="2026-03-13 09:19:11 +0000 UTC" firstStartedPulling="2026-03-13 09:19:12.649111736 +0000 UTC m=+393.399026413" lastFinishedPulling="2026-03-13 09:19:25.107212523 +0000 UTC m=+405.857127220" observedRunningTime="2026-03-13 09:19:25.853108163 +0000 UTC m=+406.603022900" watchObservedRunningTime="2026-03-13 09:19:25.857292936 +0000 UTC m=+406.607207623" Mar 13 09:19:25 crc kubenswrapper[4930]: I0313 09:19:25.897975 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/prometheus-k8s-0" podStartSLOduration=4.65871657 podStartE2EDuration="8.897955432s" podCreationTimestamp="2026-03-13 09:19:17 +0000 UTC" firstStartedPulling="2026-03-13 09:19:18.727465194 +0000 UTC m=+399.477379871" lastFinishedPulling="2026-03-13 09:19:22.966704056 +0000 UTC m=+403.716618733" observedRunningTime="2026-03-13 09:19:25.891080182 +0000 UTC m=+406.640994879" watchObservedRunningTime="2026-03-13 09:19:25.897955432 +0000 UTC m=+406.647870119" Mar 13 09:19:25 crc kubenswrapper[4930]: I0313 09:19:25.903248 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:25 crc kubenswrapper[4930]: I0313 09:19:25.903660 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:25 crc kubenswrapper[4930]: I0313 09:19:25.908366 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:26 crc kubenswrapper[4930]: I0313 09:19:26.827766 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" event={"ID":"951e59b6-d39a-4ec2-beda-4b6b8b3e20c8","Type":"ContainerStarted","Data":"462ac89bfa7f7d494bddb33241b77e3efda7d435fe9a1820ce52eeb92e87ac4d"} Mar 13 09:19:26 crc kubenswrapper[4930]: I0313 09:19:26.828281 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:26 crc kubenswrapper[4930]: I0313 09:19:26.845263 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:19:26 crc kubenswrapper[4930]: I0313 09:19:26.849302 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" Mar 13 09:19:26 crc kubenswrapper[4930]: I0313 09:19:26.887872 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" podStartSLOduration=3.8256577849999998 podStartE2EDuration="14.887854504s" podCreationTimestamp="2026-03-13 09:19:12 +0000 UTC" firstStartedPulling="2026-03-13 09:19:13.921411888 +0000 UTC m=+394.671326565" lastFinishedPulling="2026-03-13 09:19:24.983608607 +0000 UTC m=+405.733523284" observedRunningTime="2026-03-13 09:19:26.857814841 +0000 UTC m=+407.607729518" watchObservedRunningTime="2026-03-13 09:19:26.887854504 +0000 UTC m=+407.637769201" Mar 13 09:19:26 crc kubenswrapper[4930]: I0313 09:19:26.941159 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-btdqm"] Mar 13 09:19:27 crc kubenswrapper[4930]: I0313 09:19:27.413517 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:19:36 crc kubenswrapper[4930]: I0313 09:19:36.517814 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:36 crc kubenswrapper[4930]: I0313 09:19:36.518350 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:42 crc kubenswrapper[4930]: I0313 09:19:42.309096 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:19:42 crc kubenswrapper[4930]: I0313 09:19:42.309731 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:19:51 crc kubenswrapper[4930]: I0313 09:19:51.988076 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-btdqm" podUID="0805c214-0e3a-44a0-8bfe-38e8adcccde4" containerName="console" containerID="cri-o://09b323c026cfd6d4a081277c5fdd13281297e57f7b3186ed70554ccd759c673a" gracePeriod=15 Mar 13 09:19:53 crc kubenswrapper[4930]: I0313 09:19:53.062392 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-btdqm_0805c214-0e3a-44a0-8bfe-38e8adcccde4/console/0.log" Mar 13 09:19:53 crc kubenswrapper[4930]: I0313 09:19:53.062743 4930 generic.go:334] "Generic (PLEG): container finished" podID="0805c214-0e3a-44a0-8bfe-38e8adcccde4" containerID="09b323c026cfd6d4a081277c5fdd13281297e57f7b3186ed70554ccd759c673a" exitCode=2 Mar 13 09:19:53 crc kubenswrapper[4930]: I0313 09:19:53.062779 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-btdqm" event={"ID":"0805c214-0e3a-44a0-8bfe-38e8adcccde4","Type":"ContainerDied","Data":"09b323c026cfd6d4a081277c5fdd13281297e57f7b3186ed70554ccd759c673a"} Mar 13 09:19:53 crc kubenswrapper[4930]: I0313 09:19:53.940812 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-btdqm_0805c214-0e3a-44a0-8bfe-38e8adcccde4/console/0.log" Mar 13 09:19:53 crc kubenswrapper[4930]: I0313 09:19:53.941174 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.049870 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-service-ca\") pod \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.049955 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-oauth-serving-cert\") pod \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.049980 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-trusted-ca-bundle\") pod \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.050031 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2668c\" (UniqueName: \"kubernetes.io/projected/0805c214-0e3a-44a0-8bfe-38e8adcccde4-kube-api-access-2668c\") pod \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.050054 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-oauth-config\") pod \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.050120 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-config\") pod \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.050183 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-serving-cert\") pod \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\" (UID: \"0805c214-0e3a-44a0-8bfe-38e8adcccde4\") " Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.050763 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-service-ca" (OuterVolumeSpecName: "service-ca") pod "0805c214-0e3a-44a0-8bfe-38e8adcccde4" (UID: "0805c214-0e3a-44a0-8bfe-38e8adcccde4"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.050845 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "0805c214-0e3a-44a0-8bfe-38e8adcccde4" (UID: "0805c214-0e3a-44a0-8bfe-38e8adcccde4"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.050873 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-config" (OuterVolumeSpecName: "console-config") pod "0805c214-0e3a-44a0-8bfe-38e8adcccde4" (UID: "0805c214-0e3a-44a0-8bfe-38e8adcccde4"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.051014 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "0805c214-0e3a-44a0-8bfe-38e8adcccde4" (UID: "0805c214-0e3a-44a0-8bfe-38e8adcccde4"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.051594 4930 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.051631 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.051652 4930 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.051673 4930 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0805c214-0e3a-44a0-8bfe-38e8adcccde4-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.055638 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "0805c214-0e3a-44a0-8bfe-38e8adcccde4" (UID: "0805c214-0e3a-44a0-8bfe-38e8adcccde4"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.056894 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "0805c214-0e3a-44a0-8bfe-38e8adcccde4" (UID: "0805c214-0e3a-44a0-8bfe-38e8adcccde4"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.057005 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0805c214-0e3a-44a0-8bfe-38e8adcccde4-kube-api-access-2668c" (OuterVolumeSpecName: "kube-api-access-2668c") pod "0805c214-0e3a-44a0-8bfe-38e8adcccde4" (UID: "0805c214-0e3a-44a0-8bfe-38e8adcccde4"). InnerVolumeSpecName "kube-api-access-2668c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.070536 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-btdqm_0805c214-0e3a-44a0-8bfe-38e8adcccde4/console/0.log" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.070617 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-btdqm" event={"ID":"0805c214-0e3a-44a0-8bfe-38e8adcccde4","Type":"ContainerDied","Data":"5671701f6deb4c53d9a9f0b7c2b92122f6822cbe62c08a424d4da663d1002859"} Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.070666 4930 scope.go:117] "RemoveContainer" containerID="09b323c026cfd6d4a081277c5fdd13281297e57f7b3186ed70554ccd759c673a" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.070768 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-btdqm" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.108759 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-btdqm"] Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.110377 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-btdqm"] Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.154782 4930 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.154807 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2668c\" (UniqueName: \"kubernetes.io/projected/0805c214-0e3a-44a0-8bfe-38e8adcccde4-kube-api-access-2668c\") on node \"crc\" DevicePath \"\"" Mar 13 09:19:54 crc kubenswrapper[4930]: I0313 09:19:54.154817 4930 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/0805c214-0e3a-44a0-8bfe-38e8adcccde4-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:19:55 crc kubenswrapper[4930]: I0313 09:19:55.981776 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0805c214-0e3a-44a0-8bfe-38e8adcccde4" path="/var/lib/kubelet/pods/0805c214-0e3a-44a0-8bfe-38e8adcccde4/volumes" Mar 13 09:19:56 crc kubenswrapper[4930]: I0313 09:19:56.523746 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:19:56 crc kubenswrapper[4930]: I0313 09:19:56.532548 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.172513 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556560-pvjlf"] Mar 13 09:20:00 crc kubenswrapper[4930]: E0313 09:20:00.173196 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0805c214-0e3a-44a0-8bfe-38e8adcccde4" containerName="console" Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.173213 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0805c214-0e3a-44a0-8bfe-38e8adcccde4" containerName="console" Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.173348 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0805c214-0e3a-44a0-8bfe-38e8adcccde4" containerName="console" Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.173985 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556560-pvjlf" Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.177925 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.178244 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.180497 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.191494 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556560-pvjlf"] Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.349229 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nphw6\" (UniqueName: \"kubernetes.io/projected/719d2ae6-516d-40cb-82c6-d125d2ea3f8c-kube-api-access-nphw6\") pod \"auto-csr-approver-29556560-pvjlf\" (UID: \"719d2ae6-516d-40cb-82c6-d125d2ea3f8c\") " pod="openshift-infra/auto-csr-approver-29556560-pvjlf" Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.450761 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nphw6\" (UniqueName: \"kubernetes.io/projected/719d2ae6-516d-40cb-82c6-d125d2ea3f8c-kube-api-access-nphw6\") pod \"auto-csr-approver-29556560-pvjlf\" (UID: \"719d2ae6-516d-40cb-82c6-d125d2ea3f8c\") " pod="openshift-infra/auto-csr-approver-29556560-pvjlf" Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.470101 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nphw6\" (UniqueName: \"kubernetes.io/projected/719d2ae6-516d-40cb-82c6-d125d2ea3f8c-kube-api-access-nphw6\") pod \"auto-csr-approver-29556560-pvjlf\" (UID: \"719d2ae6-516d-40cb-82c6-d125d2ea3f8c\") " pod="openshift-infra/auto-csr-approver-29556560-pvjlf" Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.506300 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556560-pvjlf" Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.732522 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556560-pvjlf"] Mar 13 09:20:00 crc kubenswrapper[4930]: I0313 09:20:00.739136 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 09:20:01 crc kubenswrapper[4930]: I0313 09:20:01.122653 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556560-pvjlf" event={"ID":"719d2ae6-516d-40cb-82c6-d125d2ea3f8c","Type":"ContainerStarted","Data":"0d3d90a86c921dcd0d8412ab3b955f1980b096aa209d1c7cdea9e268e7c34443"} Mar 13 09:20:02 crc kubenswrapper[4930]: I0313 09:20:02.130532 4930 generic.go:334] "Generic (PLEG): container finished" podID="719d2ae6-516d-40cb-82c6-d125d2ea3f8c" containerID="d9db5a475b2e97fb11b342ee5ea95fed760a033eb6b79a64d15dc3e4af480f4a" exitCode=0 Mar 13 09:20:02 crc kubenswrapper[4930]: I0313 09:20:02.130726 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556560-pvjlf" event={"ID":"719d2ae6-516d-40cb-82c6-d125d2ea3f8c","Type":"ContainerDied","Data":"d9db5a475b2e97fb11b342ee5ea95fed760a033eb6b79a64d15dc3e4af480f4a"} Mar 13 09:20:03 crc kubenswrapper[4930]: I0313 09:20:03.393834 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556560-pvjlf" Mar 13 09:20:03 crc kubenswrapper[4930]: I0313 09:20:03.593799 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nphw6\" (UniqueName: \"kubernetes.io/projected/719d2ae6-516d-40cb-82c6-d125d2ea3f8c-kube-api-access-nphw6\") pod \"719d2ae6-516d-40cb-82c6-d125d2ea3f8c\" (UID: \"719d2ae6-516d-40cb-82c6-d125d2ea3f8c\") " Mar 13 09:20:03 crc kubenswrapper[4930]: I0313 09:20:03.602616 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/719d2ae6-516d-40cb-82c6-d125d2ea3f8c-kube-api-access-nphw6" (OuterVolumeSpecName: "kube-api-access-nphw6") pod "719d2ae6-516d-40cb-82c6-d125d2ea3f8c" (UID: "719d2ae6-516d-40cb-82c6-d125d2ea3f8c"). InnerVolumeSpecName "kube-api-access-nphw6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:20:03 crc kubenswrapper[4930]: I0313 09:20:03.695418 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nphw6\" (UniqueName: \"kubernetes.io/projected/719d2ae6-516d-40cb-82c6-d125d2ea3f8c-kube-api-access-nphw6\") on node \"crc\" DevicePath \"\"" Mar 13 09:20:04 crc kubenswrapper[4930]: I0313 09:20:04.144221 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556560-pvjlf" event={"ID":"719d2ae6-516d-40cb-82c6-d125d2ea3f8c","Type":"ContainerDied","Data":"0d3d90a86c921dcd0d8412ab3b955f1980b096aa209d1c7cdea9e268e7c34443"} Mar 13 09:20:04 crc kubenswrapper[4930]: I0313 09:20:04.144260 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0d3d90a86c921dcd0d8412ab3b955f1980b096aa209d1c7cdea9e268e7c34443" Mar 13 09:20:04 crc kubenswrapper[4930]: I0313 09:20:04.144301 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556560-pvjlf" Mar 13 09:20:12 crc kubenswrapper[4930]: I0313 09:20:12.308898 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:20:12 crc kubenswrapper[4930]: I0313 09:20:12.309938 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:20:12 crc kubenswrapper[4930]: I0313 09:20:12.310014 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:20:12 crc kubenswrapper[4930]: I0313 09:20:12.311273 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df84b7067d27878debd4bec4f38e693600039c3ae7e13f91f3d5b6a67bef6a54"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 09:20:12 crc kubenswrapper[4930]: I0313 09:20:12.311414 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://df84b7067d27878debd4bec4f38e693600039c3ae7e13f91f3d5b6a67bef6a54" gracePeriod=600 Mar 13 09:20:13 crc kubenswrapper[4930]: I0313 09:20:13.201659 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="df84b7067d27878debd4bec4f38e693600039c3ae7e13f91f3d5b6a67bef6a54" exitCode=0 Mar 13 09:20:13 crc kubenswrapper[4930]: I0313 09:20:13.202278 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"df84b7067d27878debd4bec4f38e693600039c3ae7e13f91f3d5b6a67bef6a54"} Mar 13 09:20:13 crc kubenswrapper[4930]: I0313 09:20:13.202305 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"8f2e2046f35c0bcc869ca63ddfbeac47d398ac629a37ba4258b97dd5da083a09"} Mar 13 09:20:13 crc kubenswrapper[4930]: I0313 09:20:13.202325 4930 scope.go:117] "RemoveContainer" containerID="75a32c3a6a592ba94cf2f17752d6cd696cc0c651141c5d595391ac31dc2f4185" Mar 13 09:20:17 crc kubenswrapper[4930]: I0313 09:20:17.413705 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:20:17 crc kubenswrapper[4930]: I0313 09:20:17.455161 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:20:18 crc kubenswrapper[4930]: I0313 09:20:18.270540 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.353237 4930 scope.go:117] "RemoveContainer" containerID="b6f73c03b04c6dfbedf2131baf5f9683ec7d3c8144448da982246657b98af284" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.412535 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-767c6dcbb4-wplpt"] Mar 13 09:20:40 crc kubenswrapper[4930]: E0313 09:20:40.413064 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="719d2ae6-516d-40cb-82c6-d125d2ea3f8c" containerName="oc" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.413083 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="719d2ae6-516d-40cb-82c6-d125d2ea3f8c" containerName="oc" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.413220 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="719d2ae6-516d-40cb-82c6-d125d2ea3f8c" containerName="oc" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.413648 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.424834 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-767c6dcbb4-wplpt"] Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.544594 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-oauth-config\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.544688 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-config\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.544707 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ddwc\" (UniqueName: \"kubernetes.io/projected/4255a8cb-2916-4af4-b84d-7c5650efb80b-kube-api-access-7ddwc\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.544793 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-serving-cert\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.544820 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-service-ca\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.544868 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-trusted-ca-bundle\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.544904 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-oauth-serving-cert\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.646188 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-config\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.646255 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ddwc\" (UniqueName: \"kubernetes.io/projected/4255a8cb-2916-4af4-b84d-7c5650efb80b-kube-api-access-7ddwc\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.646294 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-serving-cert\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.646330 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-service-ca\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.646357 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-trusted-ca-bundle\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.647226 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-config\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.647523 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-service-ca\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.647646 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-oauth-serving-cert\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.648055 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-trusted-ca-bundle\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.648371 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-oauth-serving-cert\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.648420 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-oauth-config\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.655880 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-oauth-config\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.656100 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-serving-cert\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.663252 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ddwc\" (UniqueName: \"kubernetes.io/projected/4255a8cb-2916-4af4-b84d-7c5650efb80b-kube-api-access-7ddwc\") pod \"console-767c6dcbb4-wplpt\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.745663 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:40 crc kubenswrapper[4930]: I0313 09:20:40.948093 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-767c6dcbb4-wplpt"] Mar 13 09:20:41 crc kubenswrapper[4930]: I0313 09:20:41.403802 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-767c6dcbb4-wplpt" event={"ID":"4255a8cb-2916-4af4-b84d-7c5650efb80b","Type":"ContainerStarted","Data":"7604123992f68375b2ae37007af141f5a622bd2bb74f40a7ebf748f175739864"} Mar 13 09:20:41 crc kubenswrapper[4930]: I0313 09:20:41.403855 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-767c6dcbb4-wplpt" event={"ID":"4255a8cb-2916-4af4-b84d-7c5650efb80b","Type":"ContainerStarted","Data":"ba7d89e82a6f2aedd026af5efab6112cce0af5cff44c3aec84bc72c7fb32d69c"} Mar 13 09:20:41 crc kubenswrapper[4930]: I0313 09:20:41.427329 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-767c6dcbb4-wplpt" podStartSLOduration=1.427302967 podStartE2EDuration="1.427302967s" podCreationTimestamp="2026-03-13 09:20:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:20:41.420735124 +0000 UTC m=+482.170649801" watchObservedRunningTime="2026-03-13 09:20:41.427302967 +0000 UTC m=+482.177217674" Mar 13 09:20:50 crc kubenswrapper[4930]: I0313 09:20:50.746366 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:50 crc kubenswrapper[4930]: I0313 09:20:50.747039 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:50 crc kubenswrapper[4930]: I0313 09:20:50.754185 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:51 crc kubenswrapper[4930]: I0313 09:20:51.468062 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:20:51 crc kubenswrapper[4930]: I0313 09:20:51.523492 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-568fc7fbc4-zps72"] Mar 13 09:21:16 crc kubenswrapper[4930]: I0313 09:21:16.577047 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-568fc7fbc4-zps72" podUID="80440d2f-0dac-42ca-a98a-855c116a80fd" containerName="console" containerID="cri-o://44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe" gracePeriod=15 Mar 13 09:21:16 crc kubenswrapper[4930]: I0313 09:21:16.938917 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-568fc7fbc4-zps72_80440d2f-0dac-42ca-a98a-855c116a80fd/console/0.log" Mar 13 09:21:16 crc kubenswrapper[4930]: I0313 09:21:16.939191 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.040128 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/80440d2f-0dac-42ca-a98a-855c116a80fd-console-serving-cert\") pod \"80440d2f-0dac-42ca-a98a-855c116a80fd\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.040259 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gn8w4\" (UniqueName: \"kubernetes.io/projected/80440d2f-0dac-42ca-a98a-855c116a80fd-kube-api-access-gn8w4\") pod \"80440d2f-0dac-42ca-a98a-855c116a80fd\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.040373 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-service-ca\") pod \"80440d2f-0dac-42ca-a98a-855c116a80fd\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.040467 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/80440d2f-0dac-42ca-a98a-855c116a80fd-console-oauth-config\") pod \"80440d2f-0dac-42ca-a98a-855c116a80fd\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.040551 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-console-config\") pod \"80440d2f-0dac-42ca-a98a-855c116a80fd\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.040579 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-trusted-ca-bundle\") pod \"80440d2f-0dac-42ca-a98a-855c116a80fd\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.040632 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-oauth-serving-cert\") pod \"80440d2f-0dac-42ca-a98a-855c116a80fd\" (UID: \"80440d2f-0dac-42ca-a98a-855c116a80fd\") " Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.041785 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-service-ca" (OuterVolumeSpecName: "service-ca") pod "80440d2f-0dac-42ca-a98a-855c116a80fd" (UID: "80440d2f-0dac-42ca-a98a-855c116a80fd"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.041870 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-console-config" (OuterVolumeSpecName: "console-config") pod "80440d2f-0dac-42ca-a98a-855c116a80fd" (UID: "80440d2f-0dac-42ca-a98a-855c116a80fd"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.042467 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "80440d2f-0dac-42ca-a98a-855c116a80fd" (UID: "80440d2f-0dac-42ca-a98a-855c116a80fd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.042519 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "80440d2f-0dac-42ca-a98a-855c116a80fd" (UID: "80440d2f-0dac-42ca-a98a-855c116a80fd"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.048062 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80440d2f-0dac-42ca-a98a-855c116a80fd-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "80440d2f-0dac-42ca-a98a-855c116a80fd" (UID: "80440d2f-0dac-42ca-a98a-855c116a80fd"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.048272 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/80440d2f-0dac-42ca-a98a-855c116a80fd-kube-api-access-gn8w4" (OuterVolumeSpecName: "kube-api-access-gn8w4") pod "80440d2f-0dac-42ca-a98a-855c116a80fd" (UID: "80440d2f-0dac-42ca-a98a-855c116a80fd"). InnerVolumeSpecName "kube-api-access-gn8w4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.049228 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/80440d2f-0dac-42ca-a98a-855c116a80fd-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "80440d2f-0dac-42ca-a98a-855c116a80fd" (UID: "80440d2f-0dac-42ca-a98a-855c116a80fd"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.142897 4930 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/80440d2f-0dac-42ca-a98a-855c116a80fd-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.142946 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gn8w4\" (UniqueName: \"kubernetes.io/projected/80440d2f-0dac-42ca-a98a-855c116a80fd-kube-api-access-gn8w4\") on node \"crc\" DevicePath \"\"" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.142961 4930 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.142973 4930 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/80440d2f-0dac-42ca-a98a-855c116a80fd-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.142984 4930 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.142996 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.143007 4930 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/80440d2f-0dac-42ca-a98a-855c116a80fd-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.627564 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-568fc7fbc4-zps72_80440d2f-0dac-42ca-a98a-855c116a80fd/console/0.log" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.627636 4930 generic.go:334] "Generic (PLEG): container finished" podID="80440d2f-0dac-42ca-a98a-855c116a80fd" containerID="44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe" exitCode=2 Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.627680 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-568fc7fbc4-zps72" event={"ID":"80440d2f-0dac-42ca-a98a-855c116a80fd","Type":"ContainerDied","Data":"44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe"} Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.627713 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-568fc7fbc4-zps72" event={"ID":"80440d2f-0dac-42ca-a98a-855c116a80fd","Type":"ContainerDied","Data":"5488f0b2e11e7ffe4d94877af41242fcee30f3a547619d65c73c0f7b93275e00"} Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.627739 4930 scope.go:117] "RemoveContainer" containerID="44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.627737 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-568fc7fbc4-zps72" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.645198 4930 scope.go:117] "RemoveContainer" containerID="44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe" Mar 13 09:21:17 crc kubenswrapper[4930]: E0313 09:21:17.646335 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe\": container with ID starting with 44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe not found: ID does not exist" containerID="44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.646381 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe"} err="failed to get container status \"44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe\": rpc error: code = NotFound desc = could not find container \"44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe\": container with ID starting with 44f06620816ec09718d72795fbc9fe7621ee2afc2698b77985e91bf84dfd6cfe not found: ID does not exist" Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.681512 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-568fc7fbc4-zps72"] Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.686518 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-568fc7fbc4-zps72"] Mar 13 09:21:17 crc kubenswrapper[4930]: I0313 09:21:17.984975 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="80440d2f-0dac-42ca-a98a-855c116a80fd" path="/var/lib/kubelet/pods/80440d2f-0dac-42ca-a98a-855c116a80fd/volumes" Mar 13 09:22:00 crc kubenswrapper[4930]: I0313 09:22:00.129892 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556562-k9kvz"] Mar 13 09:22:00 crc kubenswrapper[4930]: E0313 09:22:00.130723 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="80440d2f-0dac-42ca-a98a-855c116a80fd" containerName="console" Mar 13 09:22:00 crc kubenswrapper[4930]: I0313 09:22:00.130776 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="80440d2f-0dac-42ca-a98a-855c116a80fd" containerName="console" Mar 13 09:22:00 crc kubenswrapper[4930]: I0313 09:22:00.130903 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="80440d2f-0dac-42ca-a98a-855c116a80fd" containerName="console" Mar 13 09:22:00 crc kubenswrapper[4930]: I0313 09:22:00.131376 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556562-k9kvz" Mar 13 09:22:00 crc kubenswrapper[4930]: I0313 09:22:00.133277 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:22:00 crc kubenswrapper[4930]: I0313 09:22:00.133416 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:22:00 crc kubenswrapper[4930]: I0313 09:22:00.133964 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:22:00 crc kubenswrapper[4930]: I0313 09:22:00.135352 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556562-k9kvz"] Mar 13 09:22:00 crc kubenswrapper[4930]: I0313 09:22:00.330896 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s87rm\" (UniqueName: \"kubernetes.io/projected/347caee5-9072-4e3a-946c-09dcf974d068-kube-api-access-s87rm\") pod \"auto-csr-approver-29556562-k9kvz\" (UID: \"347caee5-9072-4e3a-946c-09dcf974d068\") " pod="openshift-infra/auto-csr-approver-29556562-k9kvz" Mar 13 09:22:00 crc kubenswrapper[4930]: I0313 09:22:00.432114 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s87rm\" (UniqueName: \"kubernetes.io/projected/347caee5-9072-4e3a-946c-09dcf974d068-kube-api-access-s87rm\") pod \"auto-csr-approver-29556562-k9kvz\" (UID: \"347caee5-9072-4e3a-946c-09dcf974d068\") " pod="openshift-infra/auto-csr-approver-29556562-k9kvz" Mar 13 09:22:00 crc kubenswrapper[4930]: I0313 09:22:00.456164 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s87rm\" (UniqueName: \"kubernetes.io/projected/347caee5-9072-4e3a-946c-09dcf974d068-kube-api-access-s87rm\") pod \"auto-csr-approver-29556562-k9kvz\" (UID: \"347caee5-9072-4e3a-946c-09dcf974d068\") " pod="openshift-infra/auto-csr-approver-29556562-k9kvz" Mar 13 09:22:00 crc kubenswrapper[4930]: I0313 09:22:00.755338 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556562-k9kvz" Mar 13 09:22:01 crc kubenswrapper[4930]: I0313 09:22:01.253530 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556562-k9kvz"] Mar 13 09:22:01 crc kubenswrapper[4930]: I0313 09:22:01.980218 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556562-k9kvz" event={"ID":"347caee5-9072-4e3a-946c-09dcf974d068","Type":"ContainerStarted","Data":"2887f0a3439addade0585ebbeb905bc3a143de8eaa92e4340c2d886af24d8821"} Mar 13 09:22:02 crc kubenswrapper[4930]: I0313 09:22:02.984972 4930 generic.go:334] "Generic (PLEG): container finished" podID="347caee5-9072-4e3a-946c-09dcf974d068" containerID="d9a2fddfc4f8d853db63aaa1a5552cac9402357e9ef7eadccc91d7a533b081e3" exitCode=0 Mar 13 09:22:02 crc kubenswrapper[4930]: I0313 09:22:02.985065 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556562-k9kvz" event={"ID":"347caee5-9072-4e3a-946c-09dcf974d068","Type":"ContainerDied","Data":"d9a2fddfc4f8d853db63aaa1a5552cac9402357e9ef7eadccc91d7a533b081e3"} Mar 13 09:22:04 crc kubenswrapper[4930]: I0313 09:22:04.225100 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556562-k9kvz" Mar 13 09:22:04 crc kubenswrapper[4930]: I0313 09:22:04.311873 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s87rm\" (UniqueName: \"kubernetes.io/projected/347caee5-9072-4e3a-946c-09dcf974d068-kube-api-access-s87rm\") pod \"347caee5-9072-4e3a-946c-09dcf974d068\" (UID: \"347caee5-9072-4e3a-946c-09dcf974d068\") " Mar 13 09:22:04 crc kubenswrapper[4930]: I0313 09:22:04.316741 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/347caee5-9072-4e3a-946c-09dcf974d068-kube-api-access-s87rm" (OuterVolumeSpecName: "kube-api-access-s87rm") pod "347caee5-9072-4e3a-946c-09dcf974d068" (UID: "347caee5-9072-4e3a-946c-09dcf974d068"). InnerVolumeSpecName "kube-api-access-s87rm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:22:04 crc kubenswrapper[4930]: I0313 09:22:04.413176 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s87rm\" (UniqueName: \"kubernetes.io/projected/347caee5-9072-4e3a-946c-09dcf974d068-kube-api-access-s87rm\") on node \"crc\" DevicePath \"\"" Mar 13 09:22:05 crc kubenswrapper[4930]: I0313 09:22:04.999614 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556562-k9kvz" event={"ID":"347caee5-9072-4e3a-946c-09dcf974d068","Type":"ContainerDied","Data":"2887f0a3439addade0585ebbeb905bc3a143de8eaa92e4340c2d886af24d8821"} Mar 13 09:22:05 crc kubenswrapper[4930]: I0313 09:22:05.000004 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2887f0a3439addade0585ebbeb905bc3a143de8eaa92e4340c2d886af24d8821" Mar 13 09:22:05 crc kubenswrapper[4930]: I0313 09:22:04.999647 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556562-k9kvz" Mar 13 09:22:05 crc kubenswrapper[4930]: I0313 09:22:05.305652 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556556-58mbz"] Mar 13 09:22:05 crc kubenswrapper[4930]: I0313 09:22:05.312572 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556556-58mbz"] Mar 13 09:22:05 crc kubenswrapper[4930]: I0313 09:22:05.980691 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="146468f7-456c-47bf-b142-cdce5b769152" path="/var/lib/kubelet/pods/146468f7-456c-47bf-b142-cdce5b769152/volumes" Mar 13 09:22:12 crc kubenswrapper[4930]: I0313 09:22:12.308788 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:22:12 crc kubenswrapper[4930]: I0313 09:22:12.309396 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:22:42 crc kubenswrapper[4930]: I0313 09:22:42.309301 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:22:42 crc kubenswrapper[4930]: I0313 09:22:42.309984 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:23:12 crc kubenswrapper[4930]: I0313 09:23:12.308211 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:23:12 crc kubenswrapper[4930]: I0313 09:23:12.309040 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:23:12 crc kubenswrapper[4930]: I0313 09:23:12.309109 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:23:12 crc kubenswrapper[4930]: I0313 09:23:12.310089 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8f2e2046f35c0bcc869ca63ddfbeac47d398ac629a37ba4258b97dd5da083a09"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 09:23:12 crc kubenswrapper[4930]: I0313 09:23:12.310192 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://8f2e2046f35c0bcc869ca63ddfbeac47d398ac629a37ba4258b97dd5da083a09" gracePeriod=600 Mar 13 09:23:12 crc kubenswrapper[4930]: I0313 09:23:12.484286 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="8f2e2046f35c0bcc869ca63ddfbeac47d398ac629a37ba4258b97dd5da083a09" exitCode=0 Mar 13 09:23:12 crc kubenswrapper[4930]: I0313 09:23:12.484362 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"8f2e2046f35c0bcc869ca63ddfbeac47d398ac629a37ba4258b97dd5da083a09"} Mar 13 09:23:12 crc kubenswrapper[4930]: I0313 09:23:12.484772 4930 scope.go:117] "RemoveContainer" containerID="df84b7067d27878debd4bec4f38e693600039c3ae7e13f91f3d5b6a67bef6a54" Mar 13 09:23:13 crc kubenswrapper[4930]: I0313 09:23:13.494525 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"e32155b41a79cf2a274b0699cb480a931ddd70ce47fbf5a1faf4047a087b2d60"} Mar 13 09:23:40 crc kubenswrapper[4930]: I0313 09:23:40.443679 4930 scope.go:117] "RemoveContainer" containerID="bf57aeac2857ab64e6e2f21e986893a35618c76531f76c7e2f544b8ce79dd1ef" Mar 13 09:23:40 crc kubenswrapper[4930]: I0313 09:23:40.888560 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s"] Mar 13 09:23:40 crc kubenswrapper[4930]: E0313 09:23:40.888785 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="347caee5-9072-4e3a-946c-09dcf974d068" containerName="oc" Mar 13 09:23:40 crc kubenswrapper[4930]: I0313 09:23:40.888796 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="347caee5-9072-4e3a-946c-09dcf974d068" containerName="oc" Mar 13 09:23:40 crc kubenswrapper[4930]: I0313 09:23:40.888903 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="347caee5-9072-4e3a-946c-09dcf974d068" containerName="oc" Mar 13 09:23:40 crc kubenswrapper[4930]: I0313 09:23:40.889732 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:40 crc kubenswrapper[4930]: I0313 09:23:40.891989 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 09:23:40 crc kubenswrapper[4930]: I0313 09:23:40.902986 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s"] Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.035256 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bcccfb30-4da9-4979-83f0-802a2da59f54-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s\" (UID: \"bcccfb30-4da9-4979-83f0-802a2da59f54\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.035354 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bcccfb30-4da9-4979-83f0-802a2da59f54-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s\" (UID: \"bcccfb30-4da9-4979-83f0-802a2da59f54\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.035383 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmwxq\" (UniqueName: \"kubernetes.io/projected/bcccfb30-4da9-4979-83f0-802a2da59f54-kube-api-access-xmwxq\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s\" (UID: \"bcccfb30-4da9-4979-83f0-802a2da59f54\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.136818 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bcccfb30-4da9-4979-83f0-802a2da59f54-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s\" (UID: \"bcccfb30-4da9-4979-83f0-802a2da59f54\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.136906 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmwxq\" (UniqueName: \"kubernetes.io/projected/bcccfb30-4da9-4979-83f0-802a2da59f54-kube-api-access-xmwxq\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s\" (UID: \"bcccfb30-4da9-4979-83f0-802a2da59f54\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.136966 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bcccfb30-4da9-4979-83f0-802a2da59f54-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s\" (UID: \"bcccfb30-4da9-4979-83f0-802a2da59f54\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.137805 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bcccfb30-4da9-4979-83f0-802a2da59f54-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s\" (UID: \"bcccfb30-4da9-4979-83f0-802a2da59f54\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.138013 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bcccfb30-4da9-4979-83f0-802a2da59f54-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s\" (UID: \"bcccfb30-4da9-4979-83f0-802a2da59f54\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.154845 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmwxq\" (UniqueName: \"kubernetes.io/projected/bcccfb30-4da9-4979-83f0-802a2da59f54-kube-api-access-xmwxq\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s\" (UID: \"bcccfb30-4da9-4979-83f0-802a2da59f54\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.238062 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.460813 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s"] Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.684539 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" event={"ID":"bcccfb30-4da9-4979-83f0-802a2da59f54","Type":"ContainerStarted","Data":"1b42b86f61dc1a5bd00fea053d5c0c92d1d957ca891d7518e92f8f1c378ecc4e"} Mar 13 09:23:41 crc kubenswrapper[4930]: I0313 09:23:41.684593 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" event={"ID":"bcccfb30-4da9-4979-83f0-802a2da59f54","Type":"ContainerStarted","Data":"37d036c3fec2da8922b6537bcc20e3b1344a571f9fb86917874f12354b2533fe"} Mar 13 09:23:42 crc kubenswrapper[4930]: I0313 09:23:42.695343 4930 generic.go:334] "Generic (PLEG): container finished" podID="bcccfb30-4da9-4979-83f0-802a2da59f54" containerID="1b42b86f61dc1a5bd00fea053d5c0c92d1d957ca891d7518e92f8f1c378ecc4e" exitCode=0 Mar 13 09:23:42 crc kubenswrapper[4930]: I0313 09:23:42.695937 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" event={"ID":"bcccfb30-4da9-4979-83f0-802a2da59f54","Type":"ContainerDied","Data":"1b42b86f61dc1a5bd00fea053d5c0c92d1d957ca891d7518e92f8f1c378ecc4e"} Mar 13 09:23:44 crc kubenswrapper[4930]: I0313 09:23:44.719191 4930 generic.go:334] "Generic (PLEG): container finished" podID="bcccfb30-4da9-4979-83f0-802a2da59f54" containerID="84490af1af734f0fffcf7817d078684b0bf7b52b01fd8d3e618034fa131e1b1e" exitCode=0 Mar 13 09:23:44 crc kubenswrapper[4930]: I0313 09:23:44.719245 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" event={"ID":"bcccfb30-4da9-4979-83f0-802a2da59f54","Type":"ContainerDied","Data":"84490af1af734f0fffcf7817d078684b0bf7b52b01fd8d3e618034fa131e1b1e"} Mar 13 09:23:45 crc kubenswrapper[4930]: I0313 09:23:45.727684 4930 generic.go:334] "Generic (PLEG): container finished" podID="bcccfb30-4da9-4979-83f0-802a2da59f54" containerID="6d067b95cd581857e46e0f1292efb865d04564a693bf00dcc80c1272cdb1a816" exitCode=0 Mar 13 09:23:45 crc kubenswrapper[4930]: I0313 09:23:45.727747 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" event={"ID":"bcccfb30-4da9-4979-83f0-802a2da59f54","Type":"ContainerDied","Data":"6d067b95cd581857e46e0f1292efb865d04564a693bf00dcc80c1272cdb1a816"} Mar 13 09:23:46 crc kubenswrapper[4930]: I0313 09:23:46.996338 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:47 crc kubenswrapper[4930]: I0313 09:23:47.119009 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bcccfb30-4da9-4979-83f0-802a2da59f54-bundle\") pod \"bcccfb30-4da9-4979-83f0-802a2da59f54\" (UID: \"bcccfb30-4da9-4979-83f0-802a2da59f54\") " Mar 13 09:23:47 crc kubenswrapper[4930]: I0313 09:23:47.119146 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmwxq\" (UniqueName: \"kubernetes.io/projected/bcccfb30-4da9-4979-83f0-802a2da59f54-kube-api-access-xmwxq\") pod \"bcccfb30-4da9-4979-83f0-802a2da59f54\" (UID: \"bcccfb30-4da9-4979-83f0-802a2da59f54\") " Mar 13 09:23:47 crc kubenswrapper[4930]: I0313 09:23:47.119231 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bcccfb30-4da9-4979-83f0-802a2da59f54-util\") pod \"bcccfb30-4da9-4979-83f0-802a2da59f54\" (UID: \"bcccfb30-4da9-4979-83f0-802a2da59f54\") " Mar 13 09:23:47 crc kubenswrapper[4930]: I0313 09:23:47.121890 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcccfb30-4da9-4979-83f0-802a2da59f54-bundle" (OuterVolumeSpecName: "bundle") pod "bcccfb30-4da9-4979-83f0-802a2da59f54" (UID: "bcccfb30-4da9-4979-83f0-802a2da59f54"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:23:47 crc kubenswrapper[4930]: I0313 09:23:47.127889 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bcccfb30-4da9-4979-83f0-802a2da59f54-kube-api-access-xmwxq" (OuterVolumeSpecName: "kube-api-access-xmwxq") pod "bcccfb30-4da9-4979-83f0-802a2da59f54" (UID: "bcccfb30-4da9-4979-83f0-802a2da59f54"). InnerVolumeSpecName "kube-api-access-xmwxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:23:47 crc kubenswrapper[4930]: I0313 09:23:47.221503 4930 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bcccfb30-4da9-4979-83f0-802a2da59f54-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:47 crc kubenswrapper[4930]: I0313 09:23:47.221546 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmwxq\" (UniqueName: \"kubernetes.io/projected/bcccfb30-4da9-4979-83f0-802a2da59f54-kube-api-access-xmwxq\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:47 crc kubenswrapper[4930]: I0313 09:23:47.333014 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bcccfb30-4da9-4979-83f0-802a2da59f54-util" (OuterVolumeSpecName: "util") pod "bcccfb30-4da9-4979-83f0-802a2da59f54" (UID: "bcccfb30-4da9-4979-83f0-802a2da59f54"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:23:47 crc kubenswrapper[4930]: I0313 09:23:47.424693 4930 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bcccfb30-4da9-4979-83f0-802a2da59f54-util\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:47 crc kubenswrapper[4930]: I0313 09:23:47.742227 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" event={"ID":"bcccfb30-4da9-4979-83f0-802a2da59f54","Type":"ContainerDied","Data":"37d036c3fec2da8922b6537bcc20e3b1344a571f9fb86917874f12354b2533fe"} Mar 13 09:23:47 crc kubenswrapper[4930]: I0313 09:23:47.742559 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37d036c3fec2da8922b6537bcc20e3b1344a571f9fb86917874f12354b2533fe" Mar 13 09:23:47 crc kubenswrapper[4930]: I0313 09:23:47.742292 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08zrl7s" Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.146333 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nvxdn"] Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.147551 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="ovn-controller" containerID="cri-o://d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876" gracePeriod=30 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.147627 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="nbdb" containerID="cri-o://156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d" gracePeriod=30 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.147708 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="northd" containerID="cri-o://12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a" gracePeriod=30 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.147778 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8" gracePeriod=30 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.147838 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="kube-rbac-proxy-node" containerID="cri-o://188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1" gracePeriod=30 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.147900 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="ovn-acl-logging" containerID="cri-o://f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867" gracePeriod=30 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.148072 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="sbdb" containerID="cri-o://a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049" gracePeriod=30 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.193025 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="ovnkube-controller" containerID="cri-o://af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594" gracePeriod=30 Mar 13 09:23:52 crc kubenswrapper[4930]: E0313 09:23:52.352627 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049 is running failed: container process not found" containerID="a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Mar 13 09:23:52 crc kubenswrapper[4930]: E0313 09:23:52.352735 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d is running failed: container process not found" containerID="156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Mar 13 09:23:52 crc kubenswrapper[4930]: E0313 09:23:52.353079 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049 is running failed: container process not found" containerID="a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Mar 13 09:23:52 crc kubenswrapper[4930]: E0313 09:23:52.353143 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d is running failed: container process not found" containerID="156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Mar 13 09:23:52 crc kubenswrapper[4930]: E0313 09:23:52.353413 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049 is running failed: container process not found" containerID="a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"sb\"\n"] Mar 13 09:23:52 crc kubenswrapper[4930]: E0313 09:23:52.353524 4930 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049 is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="sbdb" Mar 13 09:23:52 crc kubenswrapper[4930]: E0313 09:23:52.353592 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d is running failed: container process not found" containerID="156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d" cmd=["/bin/bash","-c","set -xeo pipefail\n. /ovnkube-lib/ovnkube-lib.sh || exit 1\novndb-readiness-probe \"nb\"\n"] Mar 13 09:23:52 crc kubenswrapper[4930]: E0313 09:23:52.353613 4930 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d is running failed: container process not found" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="nbdb" Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.781571 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b2g6h_a92650e8-d2ad-4d63-a21b-998c19841660/kube-multus/0.log" Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.781930 4930 generic.go:334] "Generic (PLEG): container finished" podID="a92650e8-d2ad-4d63-a21b-998c19841660" containerID="867a5cfb15a92126f39e08a86bc404cb589b0461be5b317f95f3058e61b4f8f9" exitCode=2 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.782010 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b2g6h" event={"ID":"a92650e8-d2ad-4d63-a21b-998c19841660","Type":"ContainerDied","Data":"867a5cfb15a92126f39e08a86bc404cb589b0461be5b317f95f3058e61b4f8f9"} Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.782655 4930 scope.go:117] "RemoveContainer" containerID="867a5cfb15a92126f39e08a86bc404cb589b0461be5b317f95f3058e61b4f8f9" Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.786753 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nvxdn_6bd55895-a75e-46de-bf2d-ed21c33217ea/ovn-acl-logging/0.log" Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787226 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nvxdn_6bd55895-a75e-46de-bf2d-ed21c33217ea/ovn-controller/0.log" Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787584 4930 generic.go:334] "Generic (PLEG): container finished" podID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerID="af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594" exitCode=0 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787606 4930 generic.go:334] "Generic (PLEG): container finished" podID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerID="a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049" exitCode=0 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787615 4930 generic.go:334] "Generic (PLEG): container finished" podID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerID="156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d" exitCode=0 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787630 4930 generic.go:334] "Generic (PLEG): container finished" podID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerID="12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a" exitCode=0 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787638 4930 generic.go:334] "Generic (PLEG): container finished" podID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerID="f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867" exitCode=143 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787651 4930 generic.go:334] "Generic (PLEG): container finished" podID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerID="d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876" exitCode=143 Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787678 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerDied","Data":"af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594"} Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787720 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerDied","Data":"a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049"} Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787738 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerDied","Data":"156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d"} Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787750 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerDied","Data":"12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a"} Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787762 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerDied","Data":"f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867"} Mar 13 09:23:52 crc kubenswrapper[4930]: I0313 09:23:52.787773 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerDied","Data":"d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876"} Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.347670 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nvxdn_6bd55895-a75e-46de-bf2d-ed21c33217ea/ovn-acl-logging/0.log" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.349167 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nvxdn_6bd55895-a75e-46de-bf2d-ed21c33217ea/ovn-controller/0.log" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.349593 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.442944 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-pgvdn"] Mar 13 09:23:53 crc kubenswrapper[4930]: E0313 09:23:53.443259 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="kubecfg-setup" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443287 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="kubecfg-setup" Mar 13 09:23:53 crc kubenswrapper[4930]: E0313 09:23:53.443301 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="ovn-controller" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443308 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="ovn-controller" Mar 13 09:23:53 crc kubenswrapper[4930]: E0313 09:23:53.443317 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="ovn-acl-logging" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443324 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="ovn-acl-logging" Mar 13 09:23:53 crc kubenswrapper[4930]: E0313 09:23:53.443336 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="sbdb" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443345 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="sbdb" Mar 13 09:23:53 crc kubenswrapper[4930]: E0313 09:23:53.443360 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcccfb30-4da9-4979-83f0-802a2da59f54" containerName="extract" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443368 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcccfb30-4da9-4979-83f0-802a2da59f54" containerName="extract" Mar 13 09:23:53 crc kubenswrapper[4930]: E0313 09:23:53.443381 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="nbdb" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443388 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="nbdb" Mar 13 09:23:53 crc kubenswrapper[4930]: E0313 09:23:53.443400 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443408 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 09:23:53 crc kubenswrapper[4930]: E0313 09:23:53.443420 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="kube-rbac-proxy-node" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443444 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="kube-rbac-proxy-node" Mar 13 09:23:53 crc kubenswrapper[4930]: E0313 09:23:53.443454 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcccfb30-4da9-4979-83f0-802a2da59f54" containerName="util" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443463 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcccfb30-4da9-4979-83f0-802a2da59f54" containerName="util" Mar 13 09:23:53 crc kubenswrapper[4930]: E0313 09:23:53.443472 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="northd" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443479 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="northd" Mar 13 09:23:53 crc kubenswrapper[4930]: E0313 09:23:53.443492 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bcccfb30-4da9-4979-83f0-802a2da59f54" containerName="pull" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443499 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bcccfb30-4da9-4979-83f0-802a2da59f54" containerName="pull" Mar 13 09:23:53 crc kubenswrapper[4930]: E0313 09:23:53.443507 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="ovnkube-controller" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443514 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="ovnkube-controller" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443646 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="sbdb" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443659 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="bcccfb30-4da9-4979-83f0-802a2da59f54" containerName="extract" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443671 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="northd" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443678 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="nbdb" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443691 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="kube-rbac-proxy-ovn-metrics" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443698 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="ovn-acl-logging" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443707 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="ovnkube-controller" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443716 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="ovn-controller" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.443727 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerName="kube-rbac-proxy-node" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.446149 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511007 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovnkube-config\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511055 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-run-netns\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511091 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-node-log\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511113 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511144 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-systemd-units\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511143 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511181 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtcxv\" (UniqueName: \"kubernetes.io/projected/6bd55895-a75e-46de-bf2d-ed21c33217ea-kube-api-access-rtcxv\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511198 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511189 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-node-log" (OuterVolumeSpecName: "node-log") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511229 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-systemd\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511273 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-openvswitch\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511304 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-env-overrides\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511261 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511345 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-etc-openvswitch\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511318 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511366 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-slash\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511421 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-cni-netd\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511424 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-slash" (OuterVolumeSpecName: "host-slash") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511464 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovnkube-script-lib\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511487 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-log-socket\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511490 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511524 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511536 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-var-lib-openvswitch\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511553 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-kubelet\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511567 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-cni-bin\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511591 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-run-ovn-kubernetes\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511613 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-ovn\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511628 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovn-node-metrics-cert\") pod \"6bd55895-a75e-46de-bf2d-ed21c33217ea\" (UID: \"6bd55895-a75e-46de-bf2d-ed21c33217ea\") " Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511796 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511814 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511845 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511869 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511898 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-log-socket" (OuterVolumeSpecName: "log-socket") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511922 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511926 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511944 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.511959 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512054 4930 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512069 4930 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512083 4930 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512094 4930 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-slash\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512104 4930 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512114 4930 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512125 4930 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-log-socket\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512137 4930 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512172 4930 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512183 4930 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512195 4930 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512206 4930 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512215 4930 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512225 4930 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512234 4930 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-node-log\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512246 4930 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.512257 4930 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.523558 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bd55895-a75e-46de-bf2d-ed21c33217ea-kube-api-access-rtcxv" (OuterVolumeSpecName: "kube-api-access-rtcxv") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "kube-api-access-rtcxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.525793 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.552907 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "6bd55895-a75e-46de-bf2d-ed21c33217ea" (UID: "6bd55895-a75e-46de-bf2d-ed21c33217ea"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.616596 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-kubelet\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.616660 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-var-lib-openvswitch\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.616694 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-log-socket\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.616742 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-systemd-units\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.616765 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-run-systemd\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.616799 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c7123547-da61-46f0-9e1c-e54b075191bb-env-overrides\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.616836 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c7123547-da61-46f0-9e1c-e54b075191bb-ovnkube-config\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.616888 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-run-ovn\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.616919 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-cni-bin\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.616968 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c7123547-da61-46f0-9e1c-e54b075191bb-ovn-node-metrics-cert\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617008 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tg27r\" (UniqueName: \"kubernetes.io/projected/c7123547-da61-46f0-9e1c-e54b075191bb-kube-api-access-tg27r\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617041 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-etc-openvswitch\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617071 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617095 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-slash\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617122 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-run-netns\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617160 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-run-ovn-kubernetes\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617193 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c7123547-da61-46f0-9e1c-e54b075191bb-ovnkube-script-lib\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617219 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-node-log\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617264 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-run-openvswitch\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617294 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-cni-netd\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617372 4930 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/6bd55895-a75e-46de-bf2d-ed21c33217ea-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617395 4930 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6bd55895-a75e-46de-bf2d-ed21c33217ea-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.617418 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtcxv\" (UniqueName: \"kubernetes.io/projected/6bd55895-a75e-46de-bf2d-ed21c33217ea-kube-api-access-rtcxv\") on node \"crc\" DevicePath \"\"" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.718657 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-run-openvswitch\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.718707 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-cni-netd\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.718767 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-kubelet\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.718800 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-run-openvswitch\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.718856 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-kubelet\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.718886 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-var-lib-openvswitch\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.718856 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-cni-netd\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.718917 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-log-socket\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.718969 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-systemd-units\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.718972 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-var-lib-openvswitch\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.718998 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-log-socket\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719011 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-run-systemd\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719035 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-systemd-units\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719079 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-run-systemd\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719041 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c7123547-da61-46f0-9e1c-e54b075191bb-env-overrides\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719131 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c7123547-da61-46f0-9e1c-e54b075191bb-ovnkube-config\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719167 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-run-ovn\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719194 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-cni-bin\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719229 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c7123547-da61-46f0-9e1c-e54b075191bb-ovn-node-metrics-cert\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719259 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-run-ovn\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719261 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tg27r\" (UniqueName: \"kubernetes.io/projected/c7123547-da61-46f0-9e1c-e54b075191bb-kube-api-access-tg27r\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719304 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-etc-openvswitch\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719330 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719351 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-slash\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719369 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-run-netns\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719395 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-run-ovn-kubernetes\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719418 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-node-log\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719469 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c7123547-da61-46f0-9e1c-e54b075191bb-ovnkube-script-lib\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719639 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-cni-bin\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719686 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c7123547-da61-46f0-9e1c-e54b075191bb-env-overrides\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719740 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-slash\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719766 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-etc-openvswitch\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719788 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719810 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-run-ovn-kubernetes\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719832 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-host-run-netns\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719856 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c7123547-da61-46f0-9e1c-e54b075191bb-node-log\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.719881 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c7123547-da61-46f0-9e1c-e54b075191bb-ovnkube-config\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.720172 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c7123547-da61-46f0-9e1c-e54b075191bb-ovnkube-script-lib\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.723619 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c7123547-da61-46f0-9e1c-e54b075191bb-ovn-node-metrics-cert\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.750940 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tg27r\" (UniqueName: \"kubernetes.io/projected/c7123547-da61-46f0-9e1c-e54b075191bb-kube-api-access-tg27r\") pod \"ovnkube-node-pgvdn\" (UID: \"c7123547-da61-46f0-9e1c-e54b075191bb\") " pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.760798 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.794152 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" event={"ID":"c7123547-da61-46f0-9e1c-e54b075191bb","Type":"ContainerStarted","Data":"7eb52abcf21a803a330fe3a331a6ab5d0b8098b1f51bb244e0bc8228d8f4849e"} Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.795847 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-b2g6h_a92650e8-d2ad-4d63-a21b-998c19841660/kube-multus/0.log" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.795898 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-b2g6h" event={"ID":"a92650e8-d2ad-4d63-a21b-998c19841660","Type":"ContainerStarted","Data":"95c271615363e0b276ae2293a64364e63f55368695b9cee24cb8da4c54b34e6a"} Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.799443 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nvxdn_6bd55895-a75e-46de-bf2d-ed21c33217ea/ovn-acl-logging/0.log" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.799979 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-nvxdn_6bd55895-a75e-46de-bf2d-ed21c33217ea/ovn-controller/0.log" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.800333 4930 generic.go:334] "Generic (PLEG): container finished" podID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerID="8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8" exitCode=0 Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.800364 4930 generic.go:334] "Generic (PLEG): container finished" podID="6bd55895-a75e-46de-bf2d-ed21c33217ea" containerID="188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1" exitCode=0 Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.800385 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerDied","Data":"8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8"} Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.800414 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerDied","Data":"188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1"} Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.800446 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" event={"ID":"6bd55895-a75e-46de-bf2d-ed21c33217ea","Type":"ContainerDied","Data":"df3e4b2d0b57c5f3f55e05e5bbda888a14f37a93cb548fad257a9e530889af86"} Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.800466 4930 scope.go:117] "RemoveContainer" containerID="af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.800613 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-nvxdn" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.836284 4930 scope.go:117] "RemoveContainer" containerID="a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.853296 4930 scope.go:117] "RemoveContainer" containerID="156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.867583 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nvxdn"] Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.873642 4930 scope.go:117] "RemoveContainer" containerID="12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.873864 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-nvxdn"] Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.903379 4930 scope.go:117] "RemoveContainer" containerID="8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.929045 4930 scope.go:117] "RemoveContainer" containerID="188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.980608 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bd55895-a75e-46de-bf2d-ed21c33217ea" path="/var/lib/kubelet/pods/6bd55895-a75e-46de-bf2d-ed21c33217ea/volumes" Mar 13 09:23:53 crc kubenswrapper[4930]: I0313 09:23:53.984591 4930 scope.go:117] "RemoveContainer" containerID="f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.001529 4930 scope.go:117] "RemoveContainer" containerID="d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.014972 4930 scope.go:117] "RemoveContainer" containerID="1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.026921 4930 scope.go:117] "RemoveContainer" containerID="af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594" Mar 13 09:23:54 crc kubenswrapper[4930]: E0313 09:23:54.027314 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594\": container with ID starting with af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594 not found: ID does not exist" containerID="af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.027346 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594"} err="failed to get container status \"af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594\": rpc error: code = NotFound desc = could not find container \"af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594\": container with ID starting with af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.027366 4930 scope.go:117] "RemoveContainer" containerID="a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049" Mar 13 09:23:54 crc kubenswrapper[4930]: E0313 09:23:54.027712 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049\": container with ID starting with a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049 not found: ID does not exist" containerID="a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.027783 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049"} err="failed to get container status \"a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049\": rpc error: code = NotFound desc = could not find container \"a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049\": container with ID starting with a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.027821 4930 scope.go:117] "RemoveContainer" containerID="156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d" Mar 13 09:23:54 crc kubenswrapper[4930]: E0313 09:23:54.028121 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d\": container with ID starting with 156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d not found: ID does not exist" containerID="156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.028146 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d"} err="failed to get container status \"156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d\": rpc error: code = NotFound desc = could not find container \"156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d\": container with ID starting with 156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.028160 4930 scope.go:117] "RemoveContainer" containerID="12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a" Mar 13 09:23:54 crc kubenswrapper[4930]: E0313 09:23:54.028383 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a\": container with ID starting with 12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a not found: ID does not exist" containerID="12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.028405 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a"} err="failed to get container status \"12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a\": rpc error: code = NotFound desc = could not find container \"12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a\": container with ID starting with 12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.028421 4930 scope.go:117] "RemoveContainer" containerID="8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8" Mar 13 09:23:54 crc kubenswrapper[4930]: E0313 09:23:54.028650 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8\": container with ID starting with 8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8 not found: ID does not exist" containerID="8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.028669 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8"} err="failed to get container status \"8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8\": rpc error: code = NotFound desc = could not find container \"8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8\": container with ID starting with 8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.028681 4930 scope.go:117] "RemoveContainer" containerID="188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1" Mar 13 09:23:54 crc kubenswrapper[4930]: E0313 09:23:54.028856 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1\": container with ID starting with 188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1 not found: ID does not exist" containerID="188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.028873 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1"} err="failed to get container status \"188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1\": rpc error: code = NotFound desc = could not find container \"188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1\": container with ID starting with 188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.028887 4930 scope.go:117] "RemoveContainer" containerID="f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867" Mar 13 09:23:54 crc kubenswrapper[4930]: E0313 09:23:54.029116 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867\": container with ID starting with f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867 not found: ID does not exist" containerID="f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.029131 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867"} err="failed to get container status \"f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867\": rpc error: code = NotFound desc = could not find container \"f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867\": container with ID starting with f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.029156 4930 scope.go:117] "RemoveContainer" containerID="d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876" Mar 13 09:23:54 crc kubenswrapper[4930]: E0313 09:23:54.029364 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876\": container with ID starting with d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876 not found: ID does not exist" containerID="d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.029383 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876"} err="failed to get container status \"d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876\": rpc error: code = NotFound desc = could not find container \"d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876\": container with ID starting with d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.029396 4930 scope.go:117] "RemoveContainer" containerID="1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486" Mar 13 09:23:54 crc kubenswrapper[4930]: E0313 09:23:54.029652 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486\": container with ID starting with 1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486 not found: ID does not exist" containerID="1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.029673 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486"} err="failed to get container status \"1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486\": rpc error: code = NotFound desc = could not find container \"1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486\": container with ID starting with 1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.029686 4930 scope.go:117] "RemoveContainer" containerID="af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.029856 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594"} err="failed to get container status \"af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594\": rpc error: code = NotFound desc = could not find container \"af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594\": container with ID starting with af4c7a25a4c0e00f975f783b04f9677968533af1605ccdd4dd22dee2f9546594 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.029873 4930 scope.go:117] "RemoveContainer" containerID="a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.030080 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049"} err="failed to get container status \"a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049\": rpc error: code = NotFound desc = could not find container \"a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049\": container with ID starting with a85f0a2828d89390e40d8b738c678ee53cf7b84a5ab11f994cccb89556d8b049 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.030096 4930 scope.go:117] "RemoveContainer" containerID="156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.030317 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d"} err="failed to get container status \"156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d\": rpc error: code = NotFound desc = could not find container \"156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d\": container with ID starting with 156ec0fff04ea29651602bba32907172f8450d1b255a4321fa5ea96b0cd0ad6d not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.030368 4930 scope.go:117] "RemoveContainer" containerID="12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.030773 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a"} err="failed to get container status \"12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a\": rpc error: code = NotFound desc = could not find container \"12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a\": container with ID starting with 12271f825c06b7a0e109ce69936303a308609623464d7e4ccdcb9cde2466f00a not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.030800 4930 scope.go:117] "RemoveContainer" containerID="8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.031037 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8"} err="failed to get container status \"8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8\": rpc error: code = NotFound desc = could not find container \"8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8\": container with ID starting with 8c9df5504949d1b8d4d96ed05cd5821111cc23c861d901d72dabc12aec2843f8 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.031058 4930 scope.go:117] "RemoveContainer" containerID="188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.031248 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1"} err="failed to get container status \"188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1\": rpc error: code = NotFound desc = could not find container \"188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1\": container with ID starting with 188c9760d0b23d3a81257a9f93fa62717667ce947d5deb643ddbb01c419d89b1 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.031266 4930 scope.go:117] "RemoveContainer" containerID="f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.031518 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867"} err="failed to get container status \"f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867\": rpc error: code = NotFound desc = could not find container \"f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867\": container with ID starting with f4a28c7df29fb164c7c1b36b86e10162a0c577b250d1ee0cbd8a6fbed7d48867 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.031543 4930 scope.go:117] "RemoveContainer" containerID="d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.031739 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876"} err="failed to get container status \"d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876\": rpc error: code = NotFound desc = could not find container \"d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876\": container with ID starting with d53663e0771d7f9e71567396c31efeec565f07292b5451ba34c8ca392f974876 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.031770 4930 scope.go:117] "RemoveContainer" containerID="1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.031922 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486"} err="failed to get container status \"1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486\": rpc error: code = NotFound desc = could not find container \"1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486\": container with ID starting with 1e1dcb4c2068763c5543856236db8e4b9215fd8d17cb05db9718bfb2fae06486 not found: ID does not exist" Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.807691 4930 generic.go:334] "Generic (PLEG): container finished" podID="c7123547-da61-46f0-9e1c-e54b075191bb" containerID="43a3ef23085250e407e62ef03725e11ef3e55563e6ab920312b6b6765c0dfc71" exitCode=0 Mar 13 09:23:54 crc kubenswrapper[4930]: I0313 09:23:54.807726 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" event={"ID":"c7123547-da61-46f0-9e1c-e54b075191bb","Type":"ContainerDied","Data":"43a3ef23085250e407e62ef03725e11ef3e55563e6ab920312b6b6765c0dfc71"} Mar 13 09:23:55 crc kubenswrapper[4930]: I0313 09:23:55.815594 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" event={"ID":"c7123547-da61-46f0-9e1c-e54b075191bb","Type":"ContainerStarted","Data":"56d8a6b992a1f37b88dbe147ea825ec0296b426f24953c936643db3695347355"} Mar 13 09:23:55 crc kubenswrapper[4930]: I0313 09:23:55.816041 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" event={"ID":"c7123547-da61-46f0-9e1c-e54b075191bb","Type":"ContainerStarted","Data":"8068e157b884950bab34ce594d6d1365271390fa8c0d1a882a79ab419937b4fb"} Mar 13 09:23:55 crc kubenswrapper[4930]: I0313 09:23:55.816051 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" event={"ID":"c7123547-da61-46f0-9e1c-e54b075191bb","Type":"ContainerStarted","Data":"4e4ab3fd1e9399108892a012bf2114e7322008aebf9bb178897a5a7e6046bd5b"} Mar 13 09:23:55 crc kubenswrapper[4930]: I0313 09:23:55.816061 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" event={"ID":"c7123547-da61-46f0-9e1c-e54b075191bb","Type":"ContainerStarted","Data":"c3e02cee60728abdb22514af0321f680c99d86f0e997f31a8f797efae18ed265"} Mar 13 09:23:55 crc kubenswrapper[4930]: I0313 09:23:55.816070 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" event={"ID":"c7123547-da61-46f0-9e1c-e54b075191bb","Type":"ContainerStarted","Data":"5e86275a2eea861fb6903ab14603c36609765588dff719e9b390ed6810414747"} Mar 13 09:23:55 crc kubenswrapper[4930]: I0313 09:23:55.816078 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" event={"ID":"c7123547-da61-46f0-9e1c-e54b075191bb","Type":"ContainerStarted","Data":"0a24985c0a9b4dfe660c58bf051dd8da68dfdcdf67bc4f7e4397724a699e58d5"} Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.485145 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv"] Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.486175 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.487908 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.488013 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-d7dw4" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.489029 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.583964 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqjgg\" (UniqueName: \"kubernetes.io/projected/b56f0bc5-964d-498b-bccf-6f7e47924a28-kube-api-access-fqjgg\") pod \"obo-prometheus-operator-68bc856cb9-jr5jv\" (UID: \"b56f0bc5-964d-498b-bccf-6f7e47924a28\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.612304 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc"] Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.612962 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.615181 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-vhkxd" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.624642 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.638109 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv"] Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.638856 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.685240 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqjgg\" (UniqueName: \"kubernetes.io/projected/b56f0bc5-964d-498b-bccf-6f7e47924a28-kube-api-access-fqjgg\") pod \"obo-prometheus-operator-68bc856cb9-jr5jv\" (UID: \"b56f0bc5-964d-498b-bccf-6f7e47924a28\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.685339 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ed9d34af-843b-4ea6-b462-048ecd45760d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc\" (UID: \"ed9d34af-843b-4ea6-b462-048ecd45760d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.685366 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ed9d34af-843b-4ea6-b462-048ecd45760d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc\" (UID: \"ed9d34af-843b-4ea6-b462-048ecd45760d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.719484 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqjgg\" (UniqueName: \"kubernetes.io/projected/b56f0bc5-964d-498b-bccf-6f7e47924a28-kube-api-access-fqjgg\") pod \"obo-prometheus-operator-68bc856cb9-jr5jv\" (UID: \"b56f0bc5-964d-498b-bccf-6f7e47924a28\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.786036 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7b88644bfd-92slv\" (UID: \"a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.786526 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ed9d34af-843b-4ea6-b462-048ecd45760d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc\" (UID: \"ed9d34af-843b-4ea6-b462-048ecd45760d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.786552 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ed9d34af-843b-4ea6-b462-048ecd45760d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc\" (UID: \"ed9d34af-843b-4ea6-b462-048ecd45760d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.786580 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7b88644bfd-92slv\" (UID: \"a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.791528 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ed9d34af-843b-4ea6-b462-048ecd45760d-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc\" (UID: \"ed9d34af-843b-4ea6-b462-048ecd45760d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.795068 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ed9d34af-843b-4ea6-b462-048ecd45760d-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc\" (UID: \"ed9d34af-843b-4ea6-b462-048ecd45760d\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.802577 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.824099 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-mdgnp"] Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.824850 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.827114 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-pw2c2" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.827325 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Mar 13 09:23:58 crc kubenswrapper[4930]: E0313 09:23:58.831825 4930 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-jr5jv_openshift-operators_b56f0bc5-964d-498b-bccf-6f7e47924a28_0(91cbfdf3bfb2a3072970562e75531e7358462e20604f817fb898f813106856f8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 09:23:58 crc kubenswrapper[4930]: E0313 09:23:58.831900 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-jr5jv_openshift-operators_b56f0bc5-964d-498b-bccf-6f7e47924a28_0(91cbfdf3bfb2a3072970562e75531e7358462e20604f817fb898f813106856f8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:23:58 crc kubenswrapper[4930]: E0313 09:23:58.831923 4930 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-jr5jv_openshift-operators_b56f0bc5-964d-498b-bccf-6f7e47924a28_0(91cbfdf3bfb2a3072970562e75531e7358462e20604f817fb898f813106856f8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:23:58 crc kubenswrapper[4930]: E0313 09:23:58.831967 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-68bc856cb9-jr5jv_openshift-operators(b56f0bc5-964d-498b-bccf-6f7e47924a28)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-68bc856cb9-jr5jv_openshift-operators(b56f0bc5-964d-498b-bccf-6f7e47924a28)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-jr5jv_openshift-operators_b56f0bc5-964d-498b-bccf-6f7e47924a28_0(91cbfdf3bfb2a3072970562e75531e7358462e20604f817fb898f813106856f8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" podUID="b56f0bc5-964d-498b-bccf-6f7e47924a28" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.841849 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" event={"ID":"c7123547-da61-46f0-9e1c-e54b075191bb","Type":"ContainerStarted","Data":"156fb209ec7d5ca3ff4dc353774efd1745cf4f9699df8dadfc11043bc348d8ff"} Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.888042 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7b88644bfd-92slv\" (UID: \"a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.888134 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7b88644bfd-92slv\" (UID: \"a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.892948 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-7b88644bfd-92slv\" (UID: \"a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.892967 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-7b88644bfd-92slv\" (UID: \"a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.929059 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:23:58 crc kubenswrapper[4930]: E0313 09:23:58.951783 4930 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc_openshift-operators_ed9d34af-843b-4ea6-b462-048ecd45760d_0(32bdfc758a3c849773bd3ccad94c6d765c959bcaeca36a51b57fbcdb7276ea69): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 09:23:58 crc kubenswrapper[4930]: E0313 09:23:58.951853 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc_openshift-operators_ed9d34af-843b-4ea6-b462-048ecd45760d_0(32bdfc758a3c849773bd3ccad94c6d765c959bcaeca36a51b57fbcdb7276ea69): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:23:58 crc kubenswrapper[4930]: E0313 09:23:58.951877 4930 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc_openshift-operators_ed9d34af-843b-4ea6-b462-048ecd45760d_0(32bdfc758a3c849773bd3ccad94c6d765c959bcaeca36a51b57fbcdb7276ea69): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:23:58 crc kubenswrapper[4930]: E0313 09:23:58.951927 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc_openshift-operators(ed9d34af-843b-4ea6-b462-048ecd45760d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc_openshift-operators(ed9d34af-843b-4ea6-b462-048ecd45760d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc_openshift-operators_ed9d34af-843b-4ea6-b462-048ecd45760d_0(32bdfc758a3c849773bd3ccad94c6d765c959bcaeca36a51b57fbcdb7276ea69): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" podUID="ed9d34af-843b-4ea6-b462-048ecd45760d" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.954184 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:23:58 crc kubenswrapper[4930]: E0313 09:23:58.976155 4930 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-92slv_openshift-operators_a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21_0(b1216f9493dfedeaa7cd0d7269accb5ce034d9713a69f560a3b53621dc4c6db2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 09:23:58 crc kubenswrapper[4930]: E0313 09:23:58.976228 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-92slv_openshift-operators_a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21_0(b1216f9493dfedeaa7cd0d7269accb5ce034d9713a69f560a3b53621dc4c6db2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:23:58 crc kubenswrapper[4930]: E0313 09:23:58.976463 4930 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-92slv_openshift-operators_a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21_0(b1216f9493dfedeaa7cd0d7269accb5ce034d9713a69f560a3b53621dc4c6db2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:23:58 crc kubenswrapper[4930]: E0313 09:23:58.976520 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7b88644bfd-92slv_openshift-operators(a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7b88644bfd-92slv_openshift-operators(a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-92slv_openshift-operators_a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21_0(b1216f9493dfedeaa7cd0d7269accb5ce034d9713a69f560a3b53621dc4c6db2): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" podUID="a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.989984 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vq7tq\" (UniqueName: \"kubernetes.io/projected/819e7015-1add-4a64-be13-d4ff9c73b228-kube-api-access-vq7tq\") pod \"observability-operator-59bdc8b94-mdgnp\" (UID: \"819e7015-1add-4a64-be13-d4ff9c73b228\") " pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:23:58 crc kubenswrapper[4930]: I0313 09:23:58.990064 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/819e7015-1add-4a64-be13-d4ff9c73b228-observability-operator-tls\") pod \"observability-operator-59bdc8b94-mdgnp\" (UID: \"819e7015-1add-4a64-be13-d4ff9c73b228\") " pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.009724 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-mbxlj"] Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.010414 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.012314 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-tnkzq" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.091097 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vq7tq\" (UniqueName: \"kubernetes.io/projected/819e7015-1add-4a64-be13-d4ff9c73b228-kube-api-access-vq7tq\") pod \"observability-operator-59bdc8b94-mdgnp\" (UID: \"819e7015-1add-4a64-be13-d4ff9c73b228\") " pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.091180 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/819e7015-1add-4a64-be13-d4ff9c73b228-observability-operator-tls\") pod \"observability-operator-59bdc8b94-mdgnp\" (UID: \"819e7015-1add-4a64-be13-d4ff9c73b228\") " pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.091226 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/58959680-d407-4688-8309-0438de9ce070-openshift-service-ca\") pod \"perses-operator-5bf474d74f-mbxlj\" (UID: \"58959680-d407-4688-8309-0438de9ce070\") " pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.091254 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njxtf\" (UniqueName: \"kubernetes.io/projected/58959680-d407-4688-8309-0438de9ce070-kube-api-access-njxtf\") pod \"perses-operator-5bf474d74f-mbxlj\" (UID: \"58959680-d407-4688-8309-0438de9ce070\") " pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.096085 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/819e7015-1add-4a64-be13-d4ff9c73b228-observability-operator-tls\") pod \"observability-operator-59bdc8b94-mdgnp\" (UID: \"819e7015-1add-4a64-be13-d4ff9c73b228\") " pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.112185 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vq7tq\" (UniqueName: \"kubernetes.io/projected/819e7015-1add-4a64-be13-d4ff9c73b228-kube-api-access-vq7tq\") pod \"observability-operator-59bdc8b94-mdgnp\" (UID: \"819e7015-1add-4a64-be13-d4ff9c73b228\") " pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.158628 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:23:59 crc kubenswrapper[4930]: E0313 09:23:59.190414 4930 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-mdgnp_openshift-operators_819e7015-1add-4a64-be13-d4ff9c73b228_0(41b4c522da7f748da2fac1dac51b1a70489a92ae02da67c270c12a0420a4fc70): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 09:23:59 crc kubenswrapper[4930]: E0313 09:23:59.190513 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-mdgnp_openshift-operators_819e7015-1add-4a64-be13-d4ff9c73b228_0(41b4c522da7f748da2fac1dac51b1a70489a92ae02da67c270c12a0420a4fc70): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:23:59 crc kubenswrapper[4930]: E0313 09:23:59.190539 4930 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-mdgnp_openshift-operators_819e7015-1add-4a64-be13-d4ff9c73b228_0(41b4c522da7f748da2fac1dac51b1a70489a92ae02da67c270c12a0420a4fc70): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:23:59 crc kubenswrapper[4930]: E0313 09:23:59.190598 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-59bdc8b94-mdgnp_openshift-operators(819e7015-1add-4a64-be13-d4ff9c73b228)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-59bdc8b94-mdgnp_openshift-operators(819e7015-1add-4a64-be13-d4ff9c73b228)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-mdgnp_openshift-operators_819e7015-1add-4a64-be13-d4ff9c73b228_0(41b4c522da7f748da2fac1dac51b1a70489a92ae02da67c270c12a0420a4fc70): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.192285 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/58959680-d407-4688-8309-0438de9ce070-openshift-service-ca\") pod \"perses-operator-5bf474d74f-mbxlj\" (UID: \"58959680-d407-4688-8309-0438de9ce070\") " pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.192363 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njxtf\" (UniqueName: \"kubernetes.io/projected/58959680-d407-4688-8309-0438de9ce070-kube-api-access-njxtf\") pod \"perses-operator-5bf474d74f-mbxlj\" (UID: \"58959680-d407-4688-8309-0438de9ce070\") " pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.193332 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/58959680-d407-4688-8309-0438de9ce070-openshift-service-ca\") pod \"perses-operator-5bf474d74f-mbxlj\" (UID: \"58959680-d407-4688-8309-0438de9ce070\") " pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.212203 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njxtf\" (UniqueName: \"kubernetes.io/projected/58959680-d407-4688-8309-0438de9ce070-kube-api-access-njxtf\") pod \"perses-operator-5bf474d74f-mbxlj\" (UID: \"58959680-d407-4688-8309-0438de9ce070\") " pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:23:59 crc kubenswrapper[4930]: I0313 09:23:59.327312 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:23:59 crc kubenswrapper[4930]: E0313 09:23:59.358366 4930 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-mbxlj_openshift-operators_58959680-d407-4688-8309-0438de9ce070_0(7dbd0cf3abdb4b5c5103617fb5066473dd86a495b6334fd8359987642e3541e0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 09:23:59 crc kubenswrapper[4930]: E0313 09:23:59.358487 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-mbxlj_openshift-operators_58959680-d407-4688-8309-0438de9ce070_0(7dbd0cf3abdb4b5c5103617fb5066473dd86a495b6334fd8359987642e3541e0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:23:59 crc kubenswrapper[4930]: E0313 09:23:59.358522 4930 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-mbxlj_openshift-operators_58959680-d407-4688-8309-0438de9ce070_0(7dbd0cf3abdb4b5c5103617fb5066473dd86a495b6334fd8359987642e3541e0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:23:59 crc kubenswrapper[4930]: E0313 09:23:59.358590 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5bf474d74f-mbxlj_openshift-operators(58959680-d407-4688-8309-0438de9ce070)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5bf474d74f-mbxlj_openshift-operators(58959680-d407-4688-8309-0438de9ce070)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-mbxlj_openshift-operators_58959680-d407-4688-8309-0438de9ce070_0(7dbd0cf3abdb4b5c5103617fb5066473dd86a495b6334fd8359987642e3541e0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" podUID="58959680-d407-4688-8309-0438de9ce070" Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.131496 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556564-rw6rl"] Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.132986 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.137072 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.137242 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.137333 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.210627 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q87vp\" (UniqueName: \"kubernetes.io/projected/b9dc22f8-83b4-4579-80b2-5b8d095acc4b-kube-api-access-q87vp\") pod \"auto-csr-approver-29556564-rw6rl\" (UID: \"b9dc22f8-83b4-4579-80b2-5b8d095acc4b\") " pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.311402 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q87vp\" (UniqueName: \"kubernetes.io/projected/b9dc22f8-83b4-4579-80b2-5b8d095acc4b-kube-api-access-q87vp\") pod \"auto-csr-approver-29556564-rw6rl\" (UID: \"b9dc22f8-83b4-4579-80b2-5b8d095acc4b\") " pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.338261 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q87vp\" (UniqueName: \"kubernetes.io/projected/b9dc22f8-83b4-4579-80b2-5b8d095acc4b-kube-api-access-q87vp\") pod \"auto-csr-approver-29556564-rw6rl\" (UID: \"b9dc22f8-83b4-4579-80b2-5b8d095acc4b\") " pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.447521 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:00 crc kubenswrapper[4930]: E0313 09:24:00.477463 4930 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29556564-rw6rl_openshift-infra_b9dc22f8-83b4-4579-80b2-5b8d095acc4b_0(e36783ec707a72973a94480b31789cd7e022838fb164c1f858b7d7fc64ffd590): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 09:24:00 crc kubenswrapper[4930]: E0313 09:24:00.477545 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29556564-rw6rl_openshift-infra_b9dc22f8-83b4-4579-80b2-5b8d095acc4b_0(e36783ec707a72973a94480b31789cd7e022838fb164c1f858b7d7fc64ffd590): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:00 crc kubenswrapper[4930]: E0313 09:24:00.477572 4930 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29556564-rw6rl_openshift-infra_b9dc22f8-83b4-4579-80b2-5b8d095acc4b_0(e36783ec707a72973a94480b31789cd7e022838fb164c1f858b7d7fc64ffd590): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:00 crc kubenswrapper[4930]: E0313 09:24:00.477631 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29556564-rw6rl_openshift-infra(b9dc22f8-83b4-4579-80b2-5b8d095acc4b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29556564-rw6rl_openshift-infra(b9dc22f8-83b4-4579-80b2-5b8d095acc4b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29556564-rw6rl_openshift-infra_b9dc22f8-83b4-4579-80b2-5b8d095acc4b_0(e36783ec707a72973a94480b31789cd7e022838fb164c1f858b7d7fc64ffd590): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" podUID="b9dc22f8-83b4-4579-80b2-5b8d095acc4b" Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.857499 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" event={"ID":"c7123547-da61-46f0-9e1c-e54b075191bb","Type":"ContainerStarted","Data":"ee35430103cb4466ae81a417e3cff1071c91a131304d84cd32c72f77410b811f"} Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.857860 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.857944 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.885874 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" podStartSLOduration=7.885855547 podStartE2EDuration="7.885855547s" podCreationTimestamp="2026-03-13 09:23:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:24:00.880513675 +0000 UTC m=+681.630428352" watchObservedRunningTime="2026-03-13 09:24:00.885855547 +0000 UTC m=+681.635770224" Mar 13 09:24:00 crc kubenswrapper[4930]: I0313 09:24:00.896222 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.392075 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc"] Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.393342 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.393903 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.407338 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-mbxlj"] Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.407490 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.407879 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.420747 4930 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc_openshift-operators_ed9d34af-843b-4ea6-b462-048ecd45760d_0(4e01b582114b51e197db43b9d69e7e6b0e672e3a5bba54c2aadf5086270e4b5d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.420812 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc_openshift-operators_ed9d34af-843b-4ea6-b462-048ecd45760d_0(4e01b582114b51e197db43b9d69e7e6b0e672e3a5bba54c2aadf5086270e4b5d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.420834 4930 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc_openshift-operators_ed9d34af-843b-4ea6-b462-048ecd45760d_0(4e01b582114b51e197db43b9d69e7e6b0e672e3a5bba54c2aadf5086270e4b5d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.420879 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc_openshift-operators(ed9d34af-843b-4ea6-b462-048ecd45760d)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc_openshift-operators(ed9d34af-843b-4ea6-b462-048ecd45760d)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc_openshift-operators_ed9d34af-843b-4ea6-b462-048ecd45760d_0(4e01b582114b51e197db43b9d69e7e6b0e672e3a5bba54c2aadf5086270e4b5d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" podUID="ed9d34af-843b-4ea6-b462-048ecd45760d" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.432685 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-mdgnp"] Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.432799 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.433274 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.438552 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv"] Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.438799 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.439310 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.442286 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556564-rw6rl"] Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.442365 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.443127 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.472875 4930 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-mbxlj_openshift-operators_58959680-d407-4688-8309-0438de9ce070_0(651a77e553964d0579237484e7b3148650d100aee7be5afaf9706b22a2a25737): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.473010 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-mbxlj_openshift-operators_58959680-d407-4688-8309-0438de9ce070_0(651a77e553964d0579237484e7b3148650d100aee7be5afaf9706b22a2a25737): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.473077 4930 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-mbxlj_openshift-operators_58959680-d407-4688-8309-0438de9ce070_0(651a77e553964d0579237484e7b3148650d100aee7be5afaf9706b22a2a25737): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.473183 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5bf474d74f-mbxlj_openshift-operators(58959680-d407-4688-8309-0438de9ce070)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5bf474d74f-mbxlj_openshift-operators(58959680-d407-4688-8309-0438de9ce070)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5bf474d74f-mbxlj_openshift-operators_58959680-d407-4688-8309-0438de9ce070_0(651a77e553964d0579237484e7b3148650d100aee7be5afaf9706b22a2a25737): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" podUID="58959680-d407-4688-8309-0438de9ce070" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.474243 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv"] Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.474345 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.474739 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.496325 4930 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-mdgnp_openshift-operators_819e7015-1add-4a64-be13-d4ff9c73b228_0(5887424f2757ba5e76336c7643f630e5a10e519d3dc98b436f1f9ab125a18257): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.496508 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-mdgnp_openshift-operators_819e7015-1add-4a64-be13-d4ff9c73b228_0(5887424f2757ba5e76336c7643f630e5a10e519d3dc98b436f1f9ab125a18257): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.496637 4930 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-mdgnp_openshift-operators_819e7015-1add-4a64-be13-d4ff9c73b228_0(5887424f2757ba5e76336c7643f630e5a10e519d3dc98b436f1f9ab125a18257): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.496768 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-59bdc8b94-mdgnp_openshift-operators(819e7015-1add-4a64-be13-d4ff9c73b228)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-59bdc8b94-mdgnp_openshift-operators(819e7015-1add-4a64-be13-d4ff9c73b228)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-59bdc8b94-mdgnp_openshift-operators_819e7015-1add-4a64-be13-d4ff9c73b228_0(5887424f2757ba5e76336c7643f630e5a10e519d3dc98b436f1f9ab125a18257): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.509335 4930 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-jr5jv_openshift-operators_b56f0bc5-964d-498b-bccf-6f7e47924a28_0(39f6ae97c5723d56eeea3014cb88cd7d3959dc4401d0b929c5156560409e426f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.509383 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-jr5jv_openshift-operators_b56f0bc5-964d-498b-bccf-6f7e47924a28_0(39f6ae97c5723d56eeea3014cb88cd7d3959dc4401d0b929c5156560409e426f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.509402 4930 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-jr5jv_openshift-operators_b56f0bc5-964d-498b-bccf-6f7e47924a28_0(39f6ae97c5723d56eeea3014cb88cd7d3959dc4401d0b929c5156560409e426f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.509454 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-68bc856cb9-jr5jv_openshift-operators(b56f0bc5-964d-498b-bccf-6f7e47924a28)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-68bc856cb9-jr5jv_openshift-operators(b56f0bc5-964d-498b-bccf-6f7e47924a28)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-68bc856cb9-jr5jv_openshift-operators_b56f0bc5-964d-498b-bccf-6f7e47924a28_0(39f6ae97c5723d56eeea3014cb88cd7d3959dc4401d0b929c5156560409e426f): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" podUID="b56f0bc5-964d-498b-bccf-6f7e47924a28" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.514202 4930 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29556564-rw6rl_openshift-infra_b9dc22f8-83b4-4579-80b2-5b8d095acc4b_0(795cc5188f61b65bee497e4b98a0d9f2b5a8dd5d755c5f4f93da9d30e688e6cf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.514323 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29556564-rw6rl_openshift-infra_b9dc22f8-83b4-4579-80b2-5b8d095acc4b_0(795cc5188f61b65bee497e4b98a0d9f2b5a8dd5d755c5f4f93da9d30e688e6cf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.514416 4930 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29556564-rw6rl_openshift-infra_b9dc22f8-83b4-4579-80b2-5b8d095acc4b_0(795cc5188f61b65bee497e4b98a0d9f2b5a8dd5d755c5f4f93da9d30e688e6cf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.514605 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29556564-rw6rl_openshift-infra(b9dc22f8-83b4-4579-80b2-5b8d095acc4b)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29556564-rw6rl_openshift-infra(b9dc22f8-83b4-4579-80b2-5b8d095acc4b)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29556564-rw6rl_openshift-infra_b9dc22f8-83b4-4579-80b2-5b8d095acc4b_0(795cc5188f61b65bee497e4b98a0d9f2b5a8dd5d755c5f4f93da9d30e688e6cf): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" podUID="b9dc22f8-83b4-4579-80b2-5b8d095acc4b" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.523316 4930 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-92slv_openshift-operators_a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21_0(53965d5cac94808116e1b3fd619c812e3b759ced5b981d1eb968480214c4eb1d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.523368 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-92slv_openshift-operators_a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21_0(53965d5cac94808116e1b3fd619c812e3b759ced5b981d1eb968480214c4eb1d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.523396 4930 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-92slv_openshift-operators_a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21_0(53965d5cac94808116e1b3fd619c812e3b759ced5b981d1eb968480214c4eb1d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:24:01 crc kubenswrapper[4930]: E0313 09:24:01.523456 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-7b88644bfd-92slv_openshift-operators(a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-7b88644bfd-92slv_openshift-operators(a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-7b88644bfd-92slv_openshift-operators_a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21_0(53965d5cac94808116e1b3fd619c812e3b759ced5b981d1eb968480214c4eb1d): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" podUID="a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.861944 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:24:01 crc kubenswrapper[4930]: I0313 09:24:01.893116 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:24:13 crc kubenswrapper[4930]: I0313 09:24:13.970636 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:24:13 crc kubenswrapper[4930]: I0313 09:24:13.971313 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:24:14 crc kubenswrapper[4930]: I0313 09:24:14.431683 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-mdgnp"] Mar 13 09:24:14 crc kubenswrapper[4930]: W0313 09:24:14.434553 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod819e7015_1add_4a64_be13_d4ff9c73b228.slice/crio-ed68009ce71eb68a7c313cf170fcb7565a0994c8fa5cad6bbd6e85b2aeb7485b WatchSource:0}: Error finding container ed68009ce71eb68a7c313cf170fcb7565a0994c8fa5cad6bbd6e85b2aeb7485b: Status 404 returned error can't find the container with id ed68009ce71eb68a7c313cf170fcb7565a0994c8fa5cad6bbd6e85b2aeb7485b Mar 13 09:24:14 crc kubenswrapper[4930]: I0313 09:24:14.945186 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" event={"ID":"819e7015-1add-4a64-be13-d4ff9c73b228","Type":"ContainerStarted","Data":"ed68009ce71eb68a7c313cf170fcb7565a0994c8fa5cad6bbd6e85b2aeb7485b"} Mar 13 09:24:14 crc kubenswrapper[4930]: I0313 09:24:14.969994 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:24:14 crc kubenswrapper[4930]: I0313 09:24:14.970753 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" Mar 13 09:24:15 crc kubenswrapper[4930]: I0313 09:24:15.268602 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv"] Mar 13 09:24:15 crc kubenswrapper[4930]: W0313 09:24:15.275701 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda9a4fc0c_e8c0_4b25_9945_bb9bc5acbf21.slice/crio-220023a59f21db543b3f5cbdfd1938a7fb8f48f53b51d9f7d28a6f0146371503 WatchSource:0}: Error finding container 220023a59f21db543b3f5cbdfd1938a7fb8f48f53b51d9f7d28a6f0146371503: Status 404 returned error can't find the container with id 220023a59f21db543b3f5cbdfd1938a7fb8f48f53b51d9f7d28a6f0146371503 Mar 13 09:24:15 crc kubenswrapper[4930]: I0313 09:24:15.953858 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" event={"ID":"a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21","Type":"ContainerStarted","Data":"220023a59f21db543b3f5cbdfd1938a7fb8f48f53b51d9f7d28a6f0146371503"} Mar 13 09:24:15 crc kubenswrapper[4930]: I0313 09:24:15.970499 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:24:15 crc kubenswrapper[4930]: I0313 09:24:15.970517 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:15 crc kubenswrapper[4930]: I0313 09:24:15.970940 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:24:15 crc kubenswrapper[4930]: I0313 09:24:15.971012 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:16 crc kubenswrapper[4930]: I0313 09:24:16.253278 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-mbxlj"] Mar 13 09:24:16 crc kubenswrapper[4930]: I0313 09:24:16.306515 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556564-rw6rl"] Mar 13 09:24:16 crc kubenswrapper[4930]: I0313 09:24:16.969894 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:24:16 crc kubenswrapper[4930]: I0313 09:24:16.969943 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:24:16 crc kubenswrapper[4930]: I0313 09:24:16.970456 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" Mar 13 09:24:16 crc kubenswrapper[4930]: I0313 09:24:16.970540 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" Mar 13 09:24:16 crc kubenswrapper[4930]: I0313 09:24:16.971208 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" event={"ID":"58959680-d407-4688-8309-0438de9ce070","Type":"ContainerStarted","Data":"6b59c1e7c1d873984cde55a4800ce0f1cb1d63d1d536af9fdfb12acae378c13b"} Mar 13 09:24:16 crc kubenswrapper[4930]: I0313 09:24:16.974639 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" event={"ID":"b9dc22f8-83b4-4579-80b2-5b8d095acc4b","Type":"ContainerStarted","Data":"e410d2f79bc433f8d016390b3ce4e2374229e4a1120bf5dd8d1ba0c75382f592"} Mar 13 09:24:17 crc kubenswrapper[4930]: I0313 09:24:17.468024 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc"] Mar 13 09:24:17 crc kubenswrapper[4930]: W0313 09:24:17.475617 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poded9d34af_843b_4ea6_b462_048ecd45760d.slice/crio-fa6ceda98cf0792f7cfb761caefe9e93ea76724861d95c885848eeb4699033e5 WatchSource:0}: Error finding container fa6ceda98cf0792f7cfb761caefe9e93ea76724861d95c885848eeb4699033e5: Status 404 returned error can't find the container with id fa6ceda98cf0792f7cfb761caefe9e93ea76724861d95c885848eeb4699033e5 Mar 13 09:24:17 crc kubenswrapper[4930]: I0313 09:24:17.546391 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv"] Mar 13 09:24:17 crc kubenswrapper[4930]: W0313 09:24:17.548358 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb56f0bc5_964d_498b_bccf_6f7e47924a28.slice/crio-e8d8c505f24cbcfc55d6f16cbd146a31947656029521ad6eeef22f480a73fbda WatchSource:0}: Error finding container e8d8c505f24cbcfc55d6f16cbd146a31947656029521ad6eeef22f480a73fbda: Status 404 returned error can't find the container with id e8d8c505f24cbcfc55d6f16cbd146a31947656029521ad6eeef22f480a73fbda Mar 13 09:24:17 crc kubenswrapper[4930]: I0313 09:24:17.980428 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" event={"ID":"ed9d34af-843b-4ea6-b462-048ecd45760d","Type":"ContainerStarted","Data":"fa6ceda98cf0792f7cfb761caefe9e93ea76724861d95c885848eeb4699033e5"} Mar 13 09:24:17 crc kubenswrapper[4930]: I0313 09:24:17.981238 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" event={"ID":"b56f0bc5-964d-498b-bccf-6f7e47924a28","Type":"ContainerStarted","Data":"e8d8c505f24cbcfc55d6f16cbd146a31947656029521ad6eeef22f480a73fbda"} Mar 13 09:24:23 crc kubenswrapper[4930]: I0313 09:24:23.796850 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.028633 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" event={"ID":"b56f0bc5-964d-498b-bccf-6f7e47924a28","Type":"ContainerStarted","Data":"420fe118e68518c602c652849203b82084f84823192bdeabed793e7a4e863686"} Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.030154 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" event={"ID":"b9dc22f8-83b4-4579-80b2-5b8d095acc4b","Type":"ContainerStarted","Data":"65aface3ab07f570391b70eada9c025d734856e880f328929b9dd4abe2dff180"} Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.031479 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" event={"ID":"819e7015-1add-4a64-be13-d4ff9c73b228","Type":"ContainerStarted","Data":"168aa9a41e863b3e1cbb2cb4acc917874653648d080c859dd1cc0524ede1dedc"} Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.031679 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.033610 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" event={"ID":"ed9d34af-843b-4ea6-b462-048ecd45760d","Type":"ContainerStarted","Data":"76e8705953db044f8c19699db7ce89f5018472d506cbc542fc34a6b3c1d867e1"} Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.033983 4930 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-mdgnp container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": dial tcp 10.217.0.13:8081: connect: connection refused" start-of-body= Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.034021 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": dial tcp 10.217.0.13:8081: connect: connection refused" Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.035924 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" event={"ID":"a9a4fc0c-e8c0-4b25-9945-bb9bc5acbf21","Type":"ContainerStarted","Data":"39b1ded88b78b355224393d7ddd984448e88ae277d375a09d65d9473c11ff1f3"} Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.037414 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" event={"ID":"58959680-d407-4688-8309-0438de9ce070","Type":"ContainerStarted","Data":"ce8c5d97f4f0755e0f45cc77e1b5b4137ccc428433b2ce5ee757cc4a896ba1ad"} Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.037550 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.042161 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-jr5jv" podStartSLOduration=20.041464836 podStartE2EDuration="26.042147126s" podCreationTimestamp="2026-03-13 09:23:58 +0000 UTC" firstStartedPulling="2026-03-13 09:24:17.551709083 +0000 UTC m=+698.301623770" lastFinishedPulling="2026-03-13 09:24:23.552391383 +0000 UTC m=+704.302306060" observedRunningTime="2026-03-13 09:24:24.041859699 +0000 UTC m=+704.791774386" watchObservedRunningTime="2026-03-13 09:24:24.042147126 +0000 UTC m=+704.792061803" Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.055066 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" podStartSLOduration=17.547415178 podStartE2EDuration="24.055046046s" podCreationTimestamp="2026-03-13 09:24:00 +0000 UTC" firstStartedPulling="2026-03-13 09:24:16.321826865 +0000 UTC m=+697.071741542" lastFinishedPulling="2026-03-13 09:24:22.829457723 +0000 UTC m=+703.579372410" observedRunningTime="2026-03-13 09:24:24.053758094 +0000 UTC m=+704.803672781" watchObservedRunningTime="2026-03-13 09:24:24.055046046 +0000 UTC m=+704.804960723" Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.082603 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-wdpdc" podStartSLOduration=19.989339264 podStartE2EDuration="26.082587278s" podCreationTimestamp="2026-03-13 09:23:58 +0000 UTC" firstStartedPulling="2026-03-13 09:24:17.47733446 +0000 UTC m=+698.227249137" lastFinishedPulling="2026-03-13 09:24:23.570582474 +0000 UTC m=+704.320497151" observedRunningTime="2026-03-13 09:24:24.079687387 +0000 UTC m=+704.829602064" watchObservedRunningTime="2026-03-13 09:24:24.082587278 +0000 UTC m=+704.832501955" Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.102889 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" podStartSLOduration=18.896806019 podStartE2EDuration="26.102872221s" podCreationTimestamp="2026-03-13 09:23:58 +0000 UTC" firstStartedPulling="2026-03-13 09:24:16.278484571 +0000 UTC m=+697.028399248" lastFinishedPulling="2026-03-13 09:24:23.484550743 +0000 UTC m=+704.234465450" observedRunningTime="2026-03-13 09:24:24.102652765 +0000 UTC m=+704.852567442" watchObservedRunningTime="2026-03-13 09:24:24.102872221 +0000 UTC m=+704.852786898" Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.127073 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podStartSLOduration=17.078459711 podStartE2EDuration="26.1270562s" podCreationTimestamp="2026-03-13 09:23:58 +0000 UTC" firstStartedPulling="2026-03-13 09:24:14.437061031 +0000 UTC m=+695.186975708" lastFinishedPulling="2026-03-13 09:24:23.48565752 +0000 UTC m=+704.235572197" observedRunningTime="2026-03-13 09:24:24.125280376 +0000 UTC m=+704.875195053" watchObservedRunningTime="2026-03-13 09:24:24.1270562 +0000 UTC m=+704.876970877" Mar 13 09:24:24 crc kubenswrapper[4930]: I0313 09:24:24.141416 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-7b88644bfd-92slv" podStartSLOduration=17.866294539 podStartE2EDuration="26.141398985s" podCreationTimestamp="2026-03-13 09:23:58 +0000 UTC" firstStartedPulling="2026-03-13 09:24:15.277314428 +0000 UTC m=+696.027229105" lastFinishedPulling="2026-03-13 09:24:23.552418874 +0000 UTC m=+704.302333551" observedRunningTime="2026-03-13 09:24:24.140529674 +0000 UTC m=+704.890444351" watchObservedRunningTime="2026-03-13 09:24:24.141398985 +0000 UTC m=+704.891313662" Mar 13 09:24:25 crc kubenswrapper[4930]: I0313 09:24:25.050086 4930 generic.go:334] "Generic (PLEG): container finished" podID="b9dc22f8-83b4-4579-80b2-5b8d095acc4b" containerID="65aface3ab07f570391b70eada9c025d734856e880f328929b9dd4abe2dff180" exitCode=0 Mar 13 09:24:25 crc kubenswrapper[4930]: I0313 09:24:25.051847 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" event={"ID":"b9dc22f8-83b4-4579-80b2-5b8d095acc4b","Type":"ContainerDied","Data":"65aface3ab07f570391b70eada9c025d734856e880f328929b9dd4abe2dff180"} Mar 13 09:24:25 crc kubenswrapper[4930]: I0313 09:24:25.059119 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 09:24:26 crc kubenswrapper[4930]: I0313 09:24:26.285707 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:26 crc kubenswrapper[4930]: I0313 09:24:26.435237 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q87vp\" (UniqueName: \"kubernetes.io/projected/b9dc22f8-83b4-4579-80b2-5b8d095acc4b-kube-api-access-q87vp\") pod \"b9dc22f8-83b4-4579-80b2-5b8d095acc4b\" (UID: \"b9dc22f8-83b4-4579-80b2-5b8d095acc4b\") " Mar 13 09:24:26 crc kubenswrapper[4930]: I0313 09:24:26.441630 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9dc22f8-83b4-4579-80b2-5b8d095acc4b-kube-api-access-q87vp" (OuterVolumeSpecName: "kube-api-access-q87vp") pod "b9dc22f8-83b4-4579-80b2-5b8d095acc4b" (UID: "b9dc22f8-83b4-4579-80b2-5b8d095acc4b"). InnerVolumeSpecName "kube-api-access-q87vp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:24:26 crc kubenswrapper[4930]: I0313 09:24:26.537276 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q87vp\" (UniqueName: \"kubernetes.io/projected/b9dc22f8-83b4-4579-80b2-5b8d095acc4b-kube-api-access-q87vp\") on node \"crc\" DevicePath \"\"" Mar 13 09:24:27 crc kubenswrapper[4930]: I0313 09:24:27.064314 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" event={"ID":"b9dc22f8-83b4-4579-80b2-5b8d095acc4b","Type":"ContainerDied","Data":"e410d2f79bc433f8d016390b3ce4e2374229e4a1120bf5dd8d1ba0c75382f592"} Mar 13 09:24:27 crc kubenswrapper[4930]: I0313 09:24:27.064361 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e410d2f79bc433f8d016390b3ce4e2374229e4a1120bf5dd8d1ba0c75382f592" Mar 13 09:24:27 crc kubenswrapper[4930]: I0313 09:24:27.064415 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556564-rw6rl" Mar 13 09:24:27 crc kubenswrapper[4930]: I0313 09:24:27.139212 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556558-j275q"] Mar 13 09:24:27 crc kubenswrapper[4930]: I0313 09:24:27.146709 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556558-j275q"] Mar 13 09:24:27 crc kubenswrapper[4930]: I0313 09:24:27.981948 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1" path="/var/lib/kubelet/pods/07ed4fe1-7bf3-4bbd-aa6a-0d37102faaf1/volumes" Mar 13 09:24:29 crc kubenswrapper[4930]: I0313 09:24:29.329888 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.606911 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-2djn9"] Mar 13 09:24:34 crc kubenswrapper[4930]: E0313 09:24:34.607585 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9dc22f8-83b4-4579-80b2-5b8d095acc4b" containerName="oc" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.607604 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9dc22f8-83b4-4579-80b2-5b8d095acc4b" containerName="oc" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.607755 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9dc22f8-83b4-4579-80b2-5b8d095acc4b" containerName="oc" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.608305 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2djn9" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.610760 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.616125 4930 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-vq256" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.616325 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.626182 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-s4m2h"] Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.627072 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-s4m2h" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.628454 4930 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-hrw6g" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.637718 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-2djn9"] Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.645929 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-s4m2h"] Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.662989 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-nnnmp"] Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.663828 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.666603 4930 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-lxzmw" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.694653 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-nnnmp"] Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.760004 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2khw\" (UniqueName: \"kubernetes.io/projected/77b779cf-5653-46d1-bbad-37b5efbbca03-kube-api-access-h2khw\") pod \"cert-manager-webhook-687f57d79b-nnnmp\" (UID: \"77b779cf-5653-46d1-bbad-37b5efbbca03\") " pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.760083 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mkrz\" (UniqueName: \"kubernetes.io/projected/6915aa10-122f-4a6b-93e3-e52ee4087eda-kube-api-access-6mkrz\") pod \"cert-manager-cainjector-cf98fcc89-2djn9\" (UID: \"6915aa10-122f-4a6b-93e3-e52ee4087eda\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-2djn9" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.760107 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssz5p\" (UniqueName: \"kubernetes.io/projected/596bf37a-3597-46b7-a155-0e809f44b02a-kube-api-access-ssz5p\") pod \"cert-manager-858654f9db-s4m2h\" (UID: \"596bf37a-3597-46b7-a155-0e809f44b02a\") " pod="cert-manager/cert-manager-858654f9db-s4m2h" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.861909 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2khw\" (UniqueName: \"kubernetes.io/projected/77b779cf-5653-46d1-bbad-37b5efbbca03-kube-api-access-h2khw\") pod \"cert-manager-webhook-687f57d79b-nnnmp\" (UID: \"77b779cf-5653-46d1-bbad-37b5efbbca03\") " pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.861981 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mkrz\" (UniqueName: \"kubernetes.io/projected/6915aa10-122f-4a6b-93e3-e52ee4087eda-kube-api-access-6mkrz\") pod \"cert-manager-cainjector-cf98fcc89-2djn9\" (UID: \"6915aa10-122f-4a6b-93e3-e52ee4087eda\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-2djn9" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.862012 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssz5p\" (UniqueName: \"kubernetes.io/projected/596bf37a-3597-46b7-a155-0e809f44b02a-kube-api-access-ssz5p\") pod \"cert-manager-858654f9db-s4m2h\" (UID: \"596bf37a-3597-46b7-a155-0e809f44b02a\") " pod="cert-manager/cert-manager-858654f9db-s4m2h" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.882939 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mkrz\" (UniqueName: \"kubernetes.io/projected/6915aa10-122f-4a6b-93e3-e52ee4087eda-kube-api-access-6mkrz\") pod \"cert-manager-cainjector-cf98fcc89-2djn9\" (UID: \"6915aa10-122f-4a6b-93e3-e52ee4087eda\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-2djn9" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.883204 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2khw\" (UniqueName: \"kubernetes.io/projected/77b779cf-5653-46d1-bbad-37b5efbbca03-kube-api-access-h2khw\") pod \"cert-manager-webhook-687f57d79b-nnnmp\" (UID: \"77b779cf-5653-46d1-bbad-37b5efbbca03\") " pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.888453 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssz5p\" (UniqueName: \"kubernetes.io/projected/596bf37a-3597-46b7-a155-0e809f44b02a-kube-api-access-ssz5p\") pod \"cert-manager-858654f9db-s4m2h\" (UID: \"596bf37a-3597-46b7-a155-0e809f44b02a\") " pod="cert-manager/cert-manager-858654f9db-s4m2h" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.929385 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2djn9" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.938591 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-s4m2h" Mar 13 09:24:34 crc kubenswrapper[4930]: I0313 09:24:34.978223 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" Mar 13 09:24:35 crc kubenswrapper[4930]: I0313 09:24:35.389770 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-2djn9"] Mar 13 09:24:35 crc kubenswrapper[4930]: W0313 09:24:35.397948 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6915aa10_122f_4a6b_93e3_e52ee4087eda.slice/crio-bad5439d7a641e2ad3e8dedcea65e98cdfecd0a1248e6df6f1f55b43a8f382d3 WatchSource:0}: Error finding container bad5439d7a641e2ad3e8dedcea65e98cdfecd0a1248e6df6f1f55b43a8f382d3: Status 404 returned error can't find the container with id bad5439d7a641e2ad3e8dedcea65e98cdfecd0a1248e6df6f1f55b43a8f382d3 Mar 13 09:24:35 crc kubenswrapper[4930]: I0313 09:24:35.398875 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-s4m2h"] Mar 13 09:24:35 crc kubenswrapper[4930]: I0313 09:24:35.449882 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-nnnmp"] Mar 13 09:24:36 crc kubenswrapper[4930]: I0313 09:24:36.118222 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-s4m2h" event={"ID":"596bf37a-3597-46b7-a155-0e809f44b02a","Type":"ContainerStarted","Data":"b325d2557407d014de923f0f3f740c5c988046cd0b1d77bd48ea1d842d34322b"} Mar 13 09:24:36 crc kubenswrapper[4930]: I0313 09:24:36.119992 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" event={"ID":"77b779cf-5653-46d1-bbad-37b5efbbca03","Type":"ContainerStarted","Data":"1d4f1a605951844db17e6ab19864bb6f6d9b9e7d8de2f01179b4d23890e8fdfb"} Mar 13 09:24:36 crc kubenswrapper[4930]: I0313 09:24:36.122121 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2djn9" event={"ID":"6915aa10-122f-4a6b-93e3-e52ee4087eda","Type":"ContainerStarted","Data":"bad5439d7a641e2ad3e8dedcea65e98cdfecd0a1248e6df6f1f55b43a8f382d3"} Mar 13 09:24:38 crc kubenswrapper[4930]: I0313 09:24:38.243864 4930 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 13 09:24:40 crc kubenswrapper[4930]: I0313 09:24:40.159796 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" event={"ID":"77b779cf-5653-46d1-bbad-37b5efbbca03","Type":"ContainerStarted","Data":"353bcdeee557ba8cb8c39791ec52b15c39ad4a3384f5fa5fa5fd8882be746bd5"} Mar 13 09:24:40 crc kubenswrapper[4930]: I0313 09:24:40.161183 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" Mar 13 09:24:40 crc kubenswrapper[4930]: I0313 09:24:40.162583 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2djn9" event={"ID":"6915aa10-122f-4a6b-93e3-e52ee4087eda","Type":"ContainerStarted","Data":"4472d178326d1d2606434f961102a3cc9ef05c5b7da35db1980169bb00e8cbd2"} Mar 13 09:24:40 crc kubenswrapper[4930]: I0313 09:24:40.164063 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-s4m2h" event={"ID":"596bf37a-3597-46b7-a155-0e809f44b02a","Type":"ContainerStarted","Data":"cb31d846bbafb8ea5419f7126a54edec31fd6ac41cae2b2be8ae940074090d52"} Mar 13 09:24:40 crc kubenswrapper[4930]: I0313 09:24:40.177517 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" podStartSLOduration=2.247393768 podStartE2EDuration="6.177490605s" podCreationTimestamp="2026-03-13 09:24:34 +0000 UTC" firstStartedPulling="2026-03-13 09:24:35.456880584 +0000 UTC m=+716.206795261" lastFinishedPulling="2026-03-13 09:24:39.386977421 +0000 UTC m=+720.136892098" observedRunningTime="2026-03-13 09:24:40.176990362 +0000 UTC m=+720.926905049" watchObservedRunningTime="2026-03-13 09:24:40.177490605 +0000 UTC m=+720.927405332" Mar 13 09:24:40 crc kubenswrapper[4930]: I0313 09:24:40.199913 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-s4m2h" podStartSLOduration=2.28617778 podStartE2EDuration="6.199889081s" podCreationTimestamp="2026-03-13 09:24:34 +0000 UTC" firstStartedPulling="2026-03-13 09:24:35.40640655 +0000 UTC m=+716.156321227" lastFinishedPulling="2026-03-13 09:24:39.320117861 +0000 UTC m=+720.070032528" observedRunningTime="2026-03-13 09:24:40.196337643 +0000 UTC m=+720.946252360" watchObservedRunningTime="2026-03-13 09:24:40.199889081 +0000 UTC m=+720.949803778" Mar 13 09:24:40 crc kubenswrapper[4930]: I0313 09:24:40.234085 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-2djn9" podStartSLOduration=3.044661618 podStartE2EDuration="6.23406291s" podCreationTimestamp="2026-03-13 09:24:34 +0000 UTC" firstStartedPulling="2026-03-13 09:24:35.400257858 +0000 UTC m=+716.150172535" lastFinishedPulling="2026-03-13 09:24:38.58965915 +0000 UTC m=+719.339573827" observedRunningTime="2026-03-13 09:24:40.230362698 +0000 UTC m=+720.980277385" watchObservedRunningTime="2026-03-13 09:24:40.23406291 +0000 UTC m=+720.983977577" Mar 13 09:24:40 crc kubenswrapper[4930]: I0313 09:24:40.500257 4930 scope.go:117] "RemoveContainer" containerID="9bd519de18d33d4bec56a1f3c7b725899af16bae93489a1d64e20c746cd894d1" Mar 13 09:24:44 crc kubenswrapper[4930]: I0313 09:24:44.982753 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.652956 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bhwmp"] Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.656712 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.674586 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhwmp"] Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.840731 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-catalog-content\") pod \"redhat-marketplace-bhwmp\" (UID: \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\") " pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.841244 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-utilities\") pod \"redhat-marketplace-bhwmp\" (UID: \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\") " pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.841336 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ns8q9\" (UniqueName: \"kubernetes.io/projected/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-kube-api-access-ns8q9\") pod \"redhat-marketplace-bhwmp\" (UID: \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\") " pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.942544 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ns8q9\" (UniqueName: \"kubernetes.io/projected/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-kube-api-access-ns8q9\") pod \"redhat-marketplace-bhwmp\" (UID: \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\") " pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.942628 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-catalog-content\") pod \"redhat-marketplace-bhwmp\" (UID: \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\") " pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.942663 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-utilities\") pod \"redhat-marketplace-bhwmp\" (UID: \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\") " pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.943099 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-utilities\") pod \"redhat-marketplace-bhwmp\" (UID: \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\") " pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.943165 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-catalog-content\") pod \"redhat-marketplace-bhwmp\" (UID: \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\") " pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.971350 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ns8q9\" (UniqueName: \"kubernetes.io/projected/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-kube-api-access-ns8q9\") pod \"redhat-marketplace-bhwmp\" (UID: \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\") " pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:24:57 crc kubenswrapper[4930]: I0313 09:24:57.978176 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:24:58 crc kubenswrapper[4930]: I0313 09:24:58.244844 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhwmp"] Mar 13 09:24:58 crc kubenswrapper[4930]: W0313 09:24:58.254679 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod553ff3e7_84eb_43cd_94e3_a34b97f04ed0.slice/crio-9b7652433fa60ab154a11472dce1899db8264f126d8dcbae70b48a3217d13b8a WatchSource:0}: Error finding container 9b7652433fa60ab154a11472dce1899db8264f126d8dcbae70b48a3217d13b8a: Status 404 returned error can't find the container with id 9b7652433fa60ab154a11472dce1899db8264f126d8dcbae70b48a3217d13b8a Mar 13 09:24:58 crc kubenswrapper[4930]: I0313 09:24:58.316661 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhwmp" event={"ID":"553ff3e7-84eb-43cd-94e3-a34b97f04ed0","Type":"ContainerStarted","Data":"9b7652433fa60ab154a11472dce1899db8264f126d8dcbae70b48a3217d13b8a"} Mar 13 09:24:59 crc kubenswrapper[4930]: I0313 09:24:59.329588 4930 generic.go:334] "Generic (PLEG): container finished" podID="553ff3e7-84eb-43cd-94e3-a34b97f04ed0" containerID="6da2fefe1ca3c02dd9bcddd2bea774144e60de0046c8341894d955050b00de17" exitCode=0 Mar 13 09:24:59 crc kubenswrapper[4930]: I0313 09:24:59.329652 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhwmp" event={"ID":"553ff3e7-84eb-43cd-94e3-a34b97f04ed0","Type":"ContainerDied","Data":"6da2fefe1ca3c02dd9bcddd2bea774144e60de0046c8341894d955050b00de17"} Mar 13 09:25:00 crc kubenswrapper[4930]: I0313 09:25:00.338739 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhwmp" event={"ID":"553ff3e7-84eb-43cd-94e3-a34b97f04ed0","Type":"ContainerStarted","Data":"26b2a3c0ce499c672037c49ea4a61905c1310def3fadf1ee4431ddd371fd6bea"} Mar 13 09:25:01 crc kubenswrapper[4930]: I0313 09:25:01.350517 4930 generic.go:334] "Generic (PLEG): container finished" podID="553ff3e7-84eb-43cd-94e3-a34b97f04ed0" containerID="26b2a3c0ce499c672037c49ea4a61905c1310def3fadf1ee4431ddd371fd6bea" exitCode=0 Mar 13 09:25:01 crc kubenswrapper[4930]: I0313 09:25:01.350619 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhwmp" event={"ID":"553ff3e7-84eb-43cd-94e3-a34b97f04ed0","Type":"ContainerDied","Data":"26b2a3c0ce499c672037c49ea4a61905c1310def3fadf1ee4431ddd371fd6bea"} Mar 13 09:25:01 crc kubenswrapper[4930]: I0313 09:25:01.356636 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.225123 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-s6wrr"] Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.234313 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.267805 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s6wrr"] Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.360453 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhwmp" event={"ID":"553ff3e7-84eb-43cd-94e3-a34b97f04ed0","Type":"ContainerStarted","Data":"495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4"} Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.377329 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bhwmp" podStartSLOduration=2.954368572 podStartE2EDuration="5.377313818s" podCreationTimestamp="2026-03-13 09:24:57 +0000 UTC" firstStartedPulling="2026-03-13 09:24:59.332665121 +0000 UTC m=+740.082579838" lastFinishedPulling="2026-03-13 09:25:01.755610407 +0000 UTC m=+742.505525084" observedRunningTime="2026-03-13 09:25:02.376604 +0000 UTC m=+743.126518677" watchObservedRunningTime="2026-03-13 09:25:02.377313818 +0000 UTC m=+743.127228495" Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.409964 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-utilities\") pod \"redhat-operators-s6wrr\" (UID: \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\") " pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.410036 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-catalog-content\") pod \"redhat-operators-s6wrr\" (UID: \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\") " pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.410060 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pvjpz\" (UniqueName: \"kubernetes.io/projected/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-kube-api-access-pvjpz\") pod \"redhat-operators-s6wrr\" (UID: \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\") " pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.510734 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-utilities\") pod \"redhat-operators-s6wrr\" (UID: \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\") " pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.510817 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-catalog-content\") pod \"redhat-operators-s6wrr\" (UID: \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\") " pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.510840 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pvjpz\" (UniqueName: \"kubernetes.io/projected/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-kube-api-access-pvjpz\") pod \"redhat-operators-s6wrr\" (UID: \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\") " pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.513062 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-catalog-content\") pod \"redhat-operators-s6wrr\" (UID: \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\") " pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.513067 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-utilities\") pod \"redhat-operators-s6wrr\" (UID: \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\") " pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.555542 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pvjpz\" (UniqueName: \"kubernetes.io/projected/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-kube-api-access-pvjpz\") pod \"redhat-operators-s6wrr\" (UID: \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\") " pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:02 crc kubenswrapper[4930]: I0313 09:25:02.560945 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:03 crc kubenswrapper[4930]: I0313 09:25:03.100569 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-s6wrr"] Mar 13 09:25:03 crc kubenswrapper[4930]: I0313 09:25:03.367826 4930 generic.go:334] "Generic (PLEG): container finished" podID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" containerID="62c3eead2924c358c13cf83d81792326f13db0fce12ab284a85af08e03d45222" exitCode=0 Mar 13 09:25:03 crc kubenswrapper[4930]: I0313 09:25:03.367945 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6wrr" event={"ID":"d62b18e7-96c4-4d13-90bb-1744e16ba9ce","Type":"ContainerDied","Data":"62c3eead2924c358c13cf83d81792326f13db0fce12ab284a85af08e03d45222"} Mar 13 09:25:03 crc kubenswrapper[4930]: I0313 09:25:03.368005 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6wrr" event={"ID":"d62b18e7-96c4-4d13-90bb-1744e16ba9ce","Type":"ContainerStarted","Data":"e60e6691139567e65742b963681d4c4271fc591f95dbbca26e530f32b51f7a75"} Mar 13 09:25:04 crc kubenswrapper[4930]: I0313 09:25:04.383978 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6wrr" event={"ID":"d62b18e7-96c4-4d13-90bb-1744e16ba9ce","Type":"ContainerStarted","Data":"9f2004a2fde40d226089cbe1ee99eb7b04568b63f482ebf50737a366343e512d"} Mar 13 09:25:05 crc kubenswrapper[4930]: I0313 09:25:05.394628 4930 generic.go:334] "Generic (PLEG): container finished" podID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" containerID="9f2004a2fde40d226089cbe1ee99eb7b04568b63f482ebf50737a366343e512d" exitCode=0 Mar 13 09:25:05 crc kubenswrapper[4930]: I0313 09:25:05.394752 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6wrr" event={"ID":"d62b18e7-96c4-4d13-90bb-1744e16ba9ce","Type":"ContainerDied","Data":"9f2004a2fde40d226089cbe1ee99eb7b04568b63f482ebf50737a366343e512d"} Mar 13 09:25:06 crc kubenswrapper[4930]: I0313 09:25:06.865950 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf"] Mar 13 09:25:06 crc kubenswrapper[4930]: I0313 09:25:06.867383 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:06 crc kubenswrapper[4930]: I0313 09:25:06.870138 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 09:25:06 crc kubenswrapper[4930]: I0313 09:25:06.884001 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-util\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf\" (UID: \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:06 crc kubenswrapper[4930]: I0313 09:25:06.884049 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-bundle\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf\" (UID: \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:06 crc kubenswrapper[4930]: I0313 09:25:06.884659 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf"] Mar 13 09:25:06 crc kubenswrapper[4930]: I0313 09:25:06.886511 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87z2n\" (UniqueName: \"kubernetes.io/projected/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-kube-api-access-87z2n\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf\" (UID: \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:06 crc kubenswrapper[4930]: I0313 09:25:06.987906 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87z2n\" (UniqueName: \"kubernetes.io/projected/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-kube-api-access-87z2n\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf\" (UID: \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:06 crc kubenswrapper[4930]: I0313 09:25:06.988322 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-util\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf\" (UID: \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:06 crc kubenswrapper[4930]: I0313 09:25:06.988352 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-bundle\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf\" (UID: \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:06 crc kubenswrapper[4930]: I0313 09:25:06.988876 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-util\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf\" (UID: \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:06 crc kubenswrapper[4930]: I0313 09:25:06.988971 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-bundle\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf\" (UID: \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.007275 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87z2n\" (UniqueName: \"kubernetes.io/projected/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-kube-api-access-87z2n\") pod \"3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf\" (UID: \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\") " pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.072424 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t"] Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.073732 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.083749 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t"] Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.090353 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9crm7\" (UniqueName: \"kubernetes.io/projected/a046e6fe-9603-44f2-a057-ce755a3b3fed-kube-api-access-9crm7\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t\" (UID: \"a046e6fe-9603-44f2-a057-ce755a3b3fed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.090498 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a046e6fe-9603-44f2-a057-ce755a3b3fed-util\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t\" (UID: \"a046e6fe-9603-44f2-a057-ce755a3b3fed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.090535 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a046e6fe-9603-44f2-a057-ce755a3b3fed-bundle\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t\" (UID: \"a046e6fe-9603-44f2-a057-ce755a3b3fed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.186401 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.192251 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9crm7\" (UniqueName: \"kubernetes.io/projected/a046e6fe-9603-44f2-a057-ce755a3b3fed-kube-api-access-9crm7\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t\" (UID: \"a046e6fe-9603-44f2-a057-ce755a3b3fed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.192325 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a046e6fe-9603-44f2-a057-ce755a3b3fed-util\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t\" (UID: \"a046e6fe-9603-44f2-a057-ce755a3b3fed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.192352 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a046e6fe-9603-44f2-a057-ce755a3b3fed-bundle\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t\" (UID: \"a046e6fe-9603-44f2-a057-ce755a3b3fed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.192841 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a046e6fe-9603-44f2-a057-ce755a3b3fed-util\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t\" (UID: \"a046e6fe-9603-44f2-a057-ce755a3b3fed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.192898 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a046e6fe-9603-44f2-a057-ce755a3b3fed-bundle\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t\" (UID: \"a046e6fe-9603-44f2-a057-ce755a3b3fed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.214136 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9crm7\" (UniqueName: \"kubernetes.io/projected/a046e6fe-9603-44f2-a057-ce755a3b3fed-kube-api-access-9crm7\") pod \"4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t\" (UID: \"a046e6fe-9603-44f2-a057-ce755a3b3fed\") " pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.373648 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf"] Mar 13 09:25:07 crc kubenswrapper[4930]: W0313 09:25:07.384835 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dc6b4ee_3be4_41b2_8fe2_b6c2fa5b078c.slice/crio-adaa699edd4da58cfdae4af62c5fe42e53e95d9b54eeccc8e0690fa9cc9be947 WatchSource:0}: Error finding container adaa699edd4da58cfdae4af62c5fe42e53e95d9b54eeccc8e0690fa9cc9be947: Status 404 returned error can't find the container with id adaa699edd4da58cfdae4af62c5fe42e53e95d9b54eeccc8e0690fa9cc9be947 Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.388496 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.418026 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6wrr" event={"ID":"d62b18e7-96c4-4d13-90bb-1744e16ba9ce","Type":"ContainerStarted","Data":"a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e"} Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.425842 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" event={"ID":"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c","Type":"ContainerStarted","Data":"adaa699edd4da58cfdae4af62c5fe42e53e95d9b54eeccc8e0690fa9cc9be947"} Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.442586 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-s6wrr" podStartSLOduration=2.121109815 podStartE2EDuration="5.442562757s" podCreationTimestamp="2026-03-13 09:25:02 +0000 UTC" firstStartedPulling="2026-03-13 09:25:03.369363616 +0000 UTC m=+744.119278293" lastFinishedPulling="2026-03-13 09:25:06.690816558 +0000 UTC m=+747.440731235" observedRunningTime="2026-03-13 09:25:07.436813214 +0000 UTC m=+748.186727901" watchObservedRunningTime="2026-03-13 09:25:07.442562757 +0000 UTC m=+748.192477434" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.594286 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t"] Mar 13 09:25:07 crc kubenswrapper[4930]: W0313 09:25:07.604519 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda046e6fe_9603_44f2_a057_ce755a3b3fed.slice/crio-306058092d10fba9c1d4bb13760e9fbea0d5d72a23b69ef9fc1c62aef54337d4 WatchSource:0}: Error finding container 306058092d10fba9c1d4bb13760e9fbea0d5d72a23b69ef9fc1c62aef54337d4: Status 404 returned error can't find the container with id 306058092d10fba9c1d4bb13760e9fbea0d5d72a23b69ef9fc1c62aef54337d4 Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.986987 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:25:07 crc kubenswrapper[4930]: I0313 09:25:07.987049 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:25:08 crc kubenswrapper[4930]: I0313 09:25:08.017009 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:25:08 crc kubenswrapper[4930]: I0313 09:25:08.434605 4930 generic.go:334] "Generic (PLEG): container finished" podID="a046e6fe-9603-44f2-a057-ce755a3b3fed" containerID="e208319674c6a22366bbe92c8e239e00aaffa2bf7fed626ab4beea2504e348ba" exitCode=0 Mar 13 09:25:08 crc kubenswrapper[4930]: I0313 09:25:08.434656 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" event={"ID":"a046e6fe-9603-44f2-a057-ce755a3b3fed","Type":"ContainerDied","Data":"e208319674c6a22366bbe92c8e239e00aaffa2bf7fed626ab4beea2504e348ba"} Mar 13 09:25:08 crc kubenswrapper[4930]: I0313 09:25:08.434717 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" event={"ID":"a046e6fe-9603-44f2-a057-ce755a3b3fed","Type":"ContainerStarted","Data":"306058092d10fba9c1d4bb13760e9fbea0d5d72a23b69ef9fc1c62aef54337d4"} Mar 13 09:25:08 crc kubenswrapper[4930]: I0313 09:25:08.436700 4930 generic.go:334] "Generic (PLEG): container finished" podID="8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" containerID="080b2df73246d3ef315a3494af1392d3c35f2d72de3fbff1c799a06272429eb9" exitCode=0 Mar 13 09:25:08 crc kubenswrapper[4930]: I0313 09:25:08.436751 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" event={"ID":"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c","Type":"ContainerDied","Data":"080b2df73246d3ef315a3494af1392d3c35f2d72de3fbff1c799a06272429eb9"} Mar 13 09:25:08 crc kubenswrapper[4930]: I0313 09:25:08.487870 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:25:10 crc kubenswrapper[4930]: I0313 09:25:10.450768 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" event={"ID":"a046e6fe-9603-44f2-a057-ce755a3b3fed","Type":"ContainerStarted","Data":"aacfa1c29644ead07c7b006634ec04b09d599ed4c45e47ad927e9b7cf792988e"} Mar 13 09:25:10 crc kubenswrapper[4930]: I0313 09:25:10.454005 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" event={"ID":"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c","Type":"ContainerStarted","Data":"de683f6918327257e894d1128037ce6ee279b3f06694aa5cdb051c8caaddf1a5"} Mar 13 09:25:11 crc kubenswrapper[4930]: I0313 09:25:11.460505 4930 generic.go:334] "Generic (PLEG): container finished" podID="8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" containerID="de683f6918327257e894d1128037ce6ee279b3f06694aa5cdb051c8caaddf1a5" exitCode=0 Mar 13 09:25:11 crc kubenswrapper[4930]: I0313 09:25:11.460570 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" event={"ID":"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c","Type":"ContainerDied","Data":"de683f6918327257e894d1128037ce6ee279b3f06694aa5cdb051c8caaddf1a5"} Mar 13 09:25:11 crc kubenswrapper[4930]: I0313 09:25:11.462413 4930 generic.go:334] "Generic (PLEG): container finished" podID="a046e6fe-9603-44f2-a057-ce755a3b3fed" containerID="aacfa1c29644ead07c7b006634ec04b09d599ed4c45e47ad927e9b7cf792988e" exitCode=0 Mar 13 09:25:11 crc kubenswrapper[4930]: I0313 09:25:11.462463 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" event={"ID":"a046e6fe-9603-44f2-a057-ce755a3b3fed","Type":"ContainerDied","Data":"aacfa1c29644ead07c7b006634ec04b09d599ed4c45e47ad927e9b7cf792988e"} Mar 13 09:25:12 crc kubenswrapper[4930]: I0313 09:25:12.308831 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:25:12 crc kubenswrapper[4930]: I0313 09:25:12.309138 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:25:12 crc kubenswrapper[4930]: I0313 09:25:12.469912 4930 generic.go:334] "Generic (PLEG): container finished" podID="a046e6fe-9603-44f2-a057-ce755a3b3fed" containerID="067415f8f8f9ecc945e879f4a920d0db0e50c2fcd7798cb7f425d4d7a100238b" exitCode=0 Mar 13 09:25:12 crc kubenswrapper[4930]: I0313 09:25:12.469952 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" event={"ID":"a046e6fe-9603-44f2-a057-ce755a3b3fed","Type":"ContainerDied","Data":"067415f8f8f9ecc945e879f4a920d0db0e50c2fcd7798cb7f425d4d7a100238b"} Mar 13 09:25:12 crc kubenswrapper[4930]: I0313 09:25:12.472017 4930 generic.go:334] "Generic (PLEG): container finished" podID="8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" containerID="725223e154f5b2dcc3989bf711f2e715898213fb853115433668224df2c48a3a" exitCode=0 Mar 13 09:25:12 crc kubenswrapper[4930]: I0313 09:25:12.472063 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" event={"ID":"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c","Type":"ContainerDied","Data":"725223e154f5b2dcc3989bf711f2e715898213fb853115433668224df2c48a3a"} Mar 13 09:25:12 crc kubenswrapper[4930]: I0313 09:25:12.562203 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:12 crc kubenswrapper[4930]: I0313 09:25:12.562530 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:12 crc kubenswrapper[4930]: I0313 09:25:12.815789 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhwmp"] Mar 13 09:25:12 crc kubenswrapper[4930]: I0313 09:25:12.816103 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bhwmp" podUID="553ff3e7-84eb-43cd-94e3-a34b97f04ed0" containerName="registry-server" containerID="cri-o://495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4" gracePeriod=2 Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.233129 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.274305 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ns8q9\" (UniqueName: \"kubernetes.io/projected/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-kube-api-access-ns8q9\") pod \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\" (UID: \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\") " Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.274352 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-utilities\") pod \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\" (UID: \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\") " Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.274452 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-catalog-content\") pod \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\" (UID: \"553ff3e7-84eb-43cd-94e3-a34b97f04ed0\") " Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.275126 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-utilities" (OuterVolumeSpecName: "utilities") pod "553ff3e7-84eb-43cd-94e3-a34b97f04ed0" (UID: "553ff3e7-84eb-43cd-94e3-a34b97f04ed0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.286406 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-kube-api-access-ns8q9" (OuterVolumeSpecName: "kube-api-access-ns8q9") pod "553ff3e7-84eb-43cd-94e3-a34b97f04ed0" (UID: "553ff3e7-84eb-43cd-94e3-a34b97f04ed0"). InnerVolumeSpecName "kube-api-access-ns8q9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.310719 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "553ff3e7-84eb-43cd-94e3-a34b97f04ed0" (UID: "553ff3e7-84eb-43cd-94e3-a34b97f04ed0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.376131 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ns8q9\" (UniqueName: \"kubernetes.io/projected/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-kube-api-access-ns8q9\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.376168 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.376178 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553ff3e7-84eb-43cd-94e3-a34b97f04ed0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.480619 4930 generic.go:334] "Generic (PLEG): container finished" podID="553ff3e7-84eb-43cd-94e3-a34b97f04ed0" containerID="495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4" exitCode=0 Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.480674 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bhwmp" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.480732 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhwmp" event={"ID":"553ff3e7-84eb-43cd-94e3-a34b97f04ed0","Type":"ContainerDied","Data":"495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4"} Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.480781 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bhwmp" event={"ID":"553ff3e7-84eb-43cd-94e3-a34b97f04ed0","Type":"ContainerDied","Data":"9b7652433fa60ab154a11472dce1899db8264f126d8dcbae70b48a3217d13b8a"} Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.480806 4930 scope.go:117] "RemoveContainer" containerID="495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.498301 4930 scope.go:117] "RemoveContainer" containerID="26b2a3c0ce499c672037c49ea4a61905c1310def3fadf1ee4431ddd371fd6bea" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.540172 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhwmp"] Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.546923 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bhwmp"] Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.557779 4930 scope.go:117] "RemoveContainer" containerID="6da2fefe1ca3c02dd9bcddd2bea774144e60de0046c8341894d955050b00de17" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.581599 4930 scope.go:117] "RemoveContainer" containerID="495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4" Mar 13 09:25:13 crc kubenswrapper[4930]: E0313 09:25:13.582245 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4\": container with ID starting with 495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4 not found: ID does not exist" containerID="495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.582295 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4"} err="failed to get container status \"495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4\": rpc error: code = NotFound desc = could not find container \"495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4\": container with ID starting with 495418ac9cd12dcec4ff4b00e18595880cc16ed43fd128f867f6f688076bf0a4 not found: ID does not exist" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.582326 4930 scope.go:117] "RemoveContainer" containerID="26b2a3c0ce499c672037c49ea4a61905c1310def3fadf1ee4431ddd371fd6bea" Mar 13 09:25:13 crc kubenswrapper[4930]: E0313 09:25:13.582934 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26b2a3c0ce499c672037c49ea4a61905c1310def3fadf1ee4431ddd371fd6bea\": container with ID starting with 26b2a3c0ce499c672037c49ea4a61905c1310def3fadf1ee4431ddd371fd6bea not found: ID does not exist" containerID="26b2a3c0ce499c672037c49ea4a61905c1310def3fadf1ee4431ddd371fd6bea" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.582980 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26b2a3c0ce499c672037c49ea4a61905c1310def3fadf1ee4431ddd371fd6bea"} err="failed to get container status \"26b2a3c0ce499c672037c49ea4a61905c1310def3fadf1ee4431ddd371fd6bea\": rpc error: code = NotFound desc = could not find container \"26b2a3c0ce499c672037c49ea4a61905c1310def3fadf1ee4431ddd371fd6bea\": container with ID starting with 26b2a3c0ce499c672037c49ea4a61905c1310def3fadf1ee4431ddd371fd6bea not found: ID does not exist" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.583014 4930 scope.go:117] "RemoveContainer" containerID="6da2fefe1ca3c02dd9bcddd2bea774144e60de0046c8341894d955050b00de17" Mar 13 09:25:13 crc kubenswrapper[4930]: E0313 09:25:13.584093 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6da2fefe1ca3c02dd9bcddd2bea774144e60de0046c8341894d955050b00de17\": container with ID starting with 6da2fefe1ca3c02dd9bcddd2bea774144e60de0046c8341894d955050b00de17 not found: ID does not exist" containerID="6da2fefe1ca3c02dd9bcddd2bea774144e60de0046c8341894d955050b00de17" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.584117 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6da2fefe1ca3c02dd9bcddd2bea774144e60de0046c8341894d955050b00de17"} err="failed to get container status \"6da2fefe1ca3c02dd9bcddd2bea774144e60de0046c8341894d955050b00de17\": rpc error: code = NotFound desc = could not find container \"6da2fefe1ca3c02dd9bcddd2bea774144e60de0046c8341894d955050b00de17\": container with ID starting with 6da2fefe1ca3c02dd9bcddd2bea774144e60de0046c8341894d955050b00de17 not found: ID does not exist" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.617789 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-s6wrr" podUID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" containerName="registry-server" probeResult="failure" output=< Mar 13 09:25:13 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:25:13 crc kubenswrapper[4930]: > Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.714691 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.773463 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.788851 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a046e6fe-9603-44f2-a057-ce755a3b3fed-bundle\") pod \"a046e6fe-9603-44f2-a057-ce755a3b3fed\" (UID: \"a046e6fe-9603-44f2-a057-ce755a3b3fed\") " Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.789284 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-util\") pod \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\" (UID: \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\") " Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.789341 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87z2n\" (UniqueName: \"kubernetes.io/projected/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-kube-api-access-87z2n\") pod \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\" (UID: \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\") " Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.789457 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9crm7\" (UniqueName: \"kubernetes.io/projected/a046e6fe-9603-44f2-a057-ce755a3b3fed-kube-api-access-9crm7\") pod \"a046e6fe-9603-44f2-a057-ce755a3b3fed\" (UID: \"a046e6fe-9603-44f2-a057-ce755a3b3fed\") " Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.789485 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-bundle\") pod \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\" (UID: \"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c\") " Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.789550 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a046e6fe-9603-44f2-a057-ce755a3b3fed-util\") pod \"a046e6fe-9603-44f2-a057-ce755a3b3fed\" (UID: \"a046e6fe-9603-44f2-a057-ce755a3b3fed\") " Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.791720 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-bundle" (OuterVolumeSpecName: "bundle") pod "8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" (UID: "8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.795777 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a046e6fe-9603-44f2-a057-ce755a3b3fed-kube-api-access-9crm7" (OuterVolumeSpecName: "kube-api-access-9crm7") pod "a046e6fe-9603-44f2-a057-ce755a3b3fed" (UID: "a046e6fe-9603-44f2-a057-ce755a3b3fed"). InnerVolumeSpecName "kube-api-access-9crm7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.796663 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-kube-api-access-87z2n" (OuterVolumeSpecName: "kube-api-access-87z2n") pod "8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" (UID: "8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c"). InnerVolumeSpecName "kube-api-access-87z2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.796717 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a046e6fe-9603-44f2-a057-ce755a3b3fed-bundle" (OuterVolumeSpecName: "bundle") pod "a046e6fe-9603-44f2-a057-ce755a3b3fed" (UID: "a046e6fe-9603-44f2-a057-ce755a3b3fed"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.865797 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-util" (OuterVolumeSpecName: "util") pod "8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" (UID: "8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.872896 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a046e6fe-9603-44f2-a057-ce755a3b3fed-util" (OuterVolumeSpecName: "util") pod "a046e6fe-9603-44f2-a057-ce755a3b3fed" (UID: "a046e6fe-9603-44f2-a057-ce755a3b3fed"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.890808 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9crm7\" (UniqueName: \"kubernetes.io/projected/a046e6fe-9603-44f2-a057-ce755a3b3fed-kube-api-access-9crm7\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.890846 4930 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.890860 4930 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a046e6fe-9603-44f2-a057-ce755a3b3fed-util\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.890869 4930 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a046e6fe-9603-44f2-a057-ce755a3b3fed-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.890877 4930 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-util\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.890888 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87z2n\" (UniqueName: \"kubernetes.io/projected/8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c-kube-api-access-87z2n\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:13 crc kubenswrapper[4930]: I0313 09:25:13.984230 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="553ff3e7-84eb-43cd-94e3-a34b97f04ed0" path="/var/lib/kubelet/pods/553ff3e7-84eb-43cd-94e3-a34b97f04ed0/volumes" Mar 13 09:25:14 crc kubenswrapper[4930]: I0313 09:25:14.494172 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" event={"ID":"8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c","Type":"ContainerDied","Data":"adaa699edd4da58cfdae4af62c5fe42e53e95d9b54eeccc8e0690fa9cc9be947"} Mar 13 09:25:14 crc kubenswrapper[4930]: I0313 09:25:14.495136 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adaa699edd4da58cfdae4af62c5fe42e53e95d9b54eeccc8e0690fa9cc9be947" Mar 13 09:25:14 crc kubenswrapper[4930]: I0313 09:25:14.494218 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/3d9a37d2dd18988fcb5ca5f4f6b82950da05d40c4031e61bc3bfef57d5jjqvf" Mar 13 09:25:14 crc kubenswrapper[4930]: I0313 09:25:14.499036 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" event={"ID":"a046e6fe-9603-44f2-a057-ce755a3b3fed","Type":"ContainerDied","Data":"306058092d10fba9c1d4bb13760e9fbea0d5d72a23b69ef9fc1c62aef54337d4"} Mar 13 09:25:14 crc kubenswrapper[4930]: I0313 09:25:14.499081 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4be416c5f2f0b2736478b7cfc76f1b991abd25af724ba21bdbdad2dd6cjbh4t" Mar 13 09:25:14 crc kubenswrapper[4930]: I0313 09:25:14.499070 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="306058092d10fba9c1d4bb13760e9fbea0d5d72a23b69ef9fc1c62aef54337d4" Mar 13 09:25:22 crc kubenswrapper[4930]: I0313 09:25:22.598635 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:22 crc kubenswrapper[4930]: I0313 09:25:22.641586 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:23 crc kubenswrapper[4930]: I0313 09:25:23.416296 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s6wrr"] Mar 13 09:25:24 crc kubenswrapper[4930]: I0313 09:25:24.564824 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-s6wrr" podUID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" containerName="registry-server" containerID="cri-o://a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e" gracePeriod=2 Mar 13 09:25:24 crc kubenswrapper[4930]: I0313 09:25:24.935033 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.040313 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-catalog-content\") pod \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\" (UID: \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\") " Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.040396 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pvjpz\" (UniqueName: \"kubernetes.io/projected/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-kube-api-access-pvjpz\") pod \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\" (UID: \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\") " Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.040479 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-utilities\") pod \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\" (UID: \"d62b18e7-96c4-4d13-90bb-1744e16ba9ce\") " Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.041128 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-utilities" (OuterVolumeSpecName: "utilities") pod "d62b18e7-96c4-4d13-90bb-1744e16ba9ce" (UID: "d62b18e7-96c4-4d13-90bb-1744e16ba9ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.053705 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-kube-api-access-pvjpz" (OuterVolumeSpecName: "kube-api-access-pvjpz") pod "d62b18e7-96c4-4d13-90bb-1744e16ba9ce" (UID: "d62b18e7-96c4-4d13-90bb-1744e16ba9ce"). InnerVolumeSpecName "kube-api-access-pvjpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.142222 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pvjpz\" (UniqueName: \"kubernetes.io/projected/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-kube-api-access-pvjpz\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.142259 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.171285 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d62b18e7-96c4-4d13-90bb-1744e16ba9ce" (UID: "d62b18e7-96c4-4d13-90bb-1744e16ba9ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.244076 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d62b18e7-96c4-4d13-90bb-1744e16ba9ce-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.574853 4930 generic.go:334] "Generic (PLEG): container finished" podID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" containerID="a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e" exitCode=0 Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.574895 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6wrr" event={"ID":"d62b18e7-96c4-4d13-90bb-1744e16ba9ce","Type":"ContainerDied","Data":"a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e"} Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.574929 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-s6wrr" event={"ID":"d62b18e7-96c4-4d13-90bb-1744e16ba9ce","Type":"ContainerDied","Data":"e60e6691139567e65742b963681d4c4271fc591f95dbbca26e530f32b51f7a75"} Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.574946 4930 scope.go:117] "RemoveContainer" containerID="a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.574952 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-s6wrr" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592030 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb"] Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.592330 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a046e6fe-9603-44f2-a057-ce755a3b3fed" containerName="util" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592343 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a046e6fe-9603-44f2-a057-ce755a3b3fed" containerName="util" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.592356 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" containerName="extract" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592362 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" containerName="extract" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.592370 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="553ff3e7-84eb-43cd-94e3-a34b97f04ed0" containerName="registry-server" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592377 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="553ff3e7-84eb-43cd-94e3-a34b97f04ed0" containerName="registry-server" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.592388 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="553ff3e7-84eb-43cd-94e3-a34b97f04ed0" containerName="extract-content" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592394 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="553ff3e7-84eb-43cd-94e3-a34b97f04ed0" containerName="extract-content" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.592402 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a046e6fe-9603-44f2-a057-ce755a3b3fed" containerName="pull" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592408 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a046e6fe-9603-44f2-a057-ce755a3b3fed" containerName="pull" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.592416 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" containerName="util" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592422 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" containerName="util" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.592447 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a046e6fe-9603-44f2-a057-ce755a3b3fed" containerName="extract" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592455 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a046e6fe-9603-44f2-a057-ce755a3b3fed" containerName="extract" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.592464 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" containerName="registry-server" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592470 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" containerName="registry-server" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.592479 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" containerName="extract-utilities" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592485 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" containerName="extract-utilities" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.592495 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" containerName="extract-content" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592500 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" containerName="extract-content" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.592507 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" containerName="pull" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592512 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" containerName="pull" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.592524 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="553ff3e7-84eb-43cd-94e3-a34b97f04ed0" containerName="extract-utilities" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592530 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="553ff3e7-84eb-43cd-94e3-a34b97f04ed0" containerName="extract-utilities" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592638 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="553ff3e7-84eb-43cd-94e3-a34b97f04ed0" containerName="registry-server" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592647 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="8dc6b4ee-3be4-41b2-8fe2-b6c2fa5b078c" containerName="extract" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592660 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" containerName="registry-server" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.592671 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="a046e6fe-9603-44f2-a057-ce755a3b3fed" containerName="extract" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.593310 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.595999 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.596054 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.596243 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.597158 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.597210 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.599300 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-kj7j7" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.600151 4930 scope.go:117] "RemoveContainer" containerID="9f2004a2fde40d226089cbe1ee99eb7b04568b63f482ebf50737a366343e512d" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.616789 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-s6wrr"] Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.631508 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-s6wrr"] Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.642210 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb"] Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.643960 4930 scope.go:117] "RemoveContainer" containerID="62c3eead2924c358c13cf83d81792326f13db0fce12ab284a85af08e03d45222" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.649162 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c17ab33f-40db-4679-8e37-e3c4d013ba29-webhook-cert\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.649447 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/c17ab33f-40db-4679-8e37-e3c4d013ba29-manager-config\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.649486 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c17ab33f-40db-4679-8e37-e3c4d013ba29-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.649511 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c17ab33f-40db-4679-8e37-e3c4d013ba29-apiservice-cert\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.649553 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jmf87\" (UniqueName: \"kubernetes.io/projected/c17ab33f-40db-4679-8e37-e3c4d013ba29-kube-api-access-jmf87\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.685303 4930 scope.go:117] "RemoveContainer" containerID="a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.685893 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e\": container with ID starting with a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e not found: ID does not exist" containerID="a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.685925 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e"} err="failed to get container status \"a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e\": rpc error: code = NotFound desc = could not find container \"a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e\": container with ID starting with a8b0ad0793ed10b9947cd16927b6b0e45345388be128615c09d2b472e67d8a9e not found: ID does not exist" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.685954 4930 scope.go:117] "RemoveContainer" containerID="9f2004a2fde40d226089cbe1ee99eb7b04568b63f482ebf50737a366343e512d" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.686299 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f2004a2fde40d226089cbe1ee99eb7b04568b63f482ebf50737a366343e512d\": container with ID starting with 9f2004a2fde40d226089cbe1ee99eb7b04568b63f482ebf50737a366343e512d not found: ID does not exist" containerID="9f2004a2fde40d226089cbe1ee99eb7b04568b63f482ebf50737a366343e512d" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.686382 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f2004a2fde40d226089cbe1ee99eb7b04568b63f482ebf50737a366343e512d"} err="failed to get container status \"9f2004a2fde40d226089cbe1ee99eb7b04568b63f482ebf50737a366343e512d\": rpc error: code = NotFound desc = could not find container \"9f2004a2fde40d226089cbe1ee99eb7b04568b63f482ebf50737a366343e512d\": container with ID starting with 9f2004a2fde40d226089cbe1ee99eb7b04568b63f482ebf50737a366343e512d not found: ID does not exist" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.686411 4930 scope.go:117] "RemoveContainer" containerID="62c3eead2924c358c13cf83d81792326f13db0fce12ab284a85af08e03d45222" Mar 13 09:25:25 crc kubenswrapper[4930]: E0313 09:25:25.687232 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62c3eead2924c358c13cf83d81792326f13db0fce12ab284a85af08e03d45222\": container with ID starting with 62c3eead2924c358c13cf83d81792326f13db0fce12ab284a85af08e03d45222 not found: ID does not exist" containerID="62c3eead2924c358c13cf83d81792326f13db0fce12ab284a85af08e03d45222" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.687267 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62c3eead2924c358c13cf83d81792326f13db0fce12ab284a85af08e03d45222"} err="failed to get container status \"62c3eead2924c358c13cf83d81792326f13db0fce12ab284a85af08e03d45222\": rpc error: code = NotFound desc = could not find container \"62c3eead2924c358c13cf83d81792326f13db0fce12ab284a85af08e03d45222\": container with ID starting with 62c3eead2924c358c13cf83d81792326f13db0fce12ab284a85af08e03d45222 not found: ID does not exist" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.727970 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-66689c4bbf-f9skd"] Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.728819 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-66689c4bbf-f9skd" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.730575 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.730907 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.731076 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-5vx28" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.750932 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jmf87\" (UniqueName: \"kubernetes.io/projected/c17ab33f-40db-4679-8e37-e3c4d013ba29-kube-api-access-jmf87\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.750982 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvfwj\" (UniqueName: \"kubernetes.io/projected/2778b5ad-9af6-40d9-a6a6-3e4f9c802c7d-kube-api-access-hvfwj\") pod \"cluster-logging-operator-66689c4bbf-f9skd\" (UID: \"2778b5ad-9af6-40d9-a6a6-3e4f9c802c7d\") " pod="openshift-logging/cluster-logging-operator-66689c4bbf-f9skd" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.751036 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c17ab33f-40db-4679-8e37-e3c4d013ba29-webhook-cert\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.751057 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/c17ab33f-40db-4679-8e37-e3c4d013ba29-manager-config\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.751082 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c17ab33f-40db-4679-8e37-e3c4d013ba29-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.751102 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c17ab33f-40db-4679-8e37-e3c4d013ba29-apiservice-cert\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.752394 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-66689c4bbf-f9skd"] Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.754373 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/c17ab33f-40db-4679-8e37-e3c4d013ba29-manager-config\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.759660 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c17ab33f-40db-4679-8e37-e3c4d013ba29-webhook-cert\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.763254 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c17ab33f-40db-4679-8e37-e3c4d013ba29-apiservice-cert\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.765237 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c17ab33f-40db-4679-8e37-e3c4d013ba29-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.776168 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jmf87\" (UniqueName: \"kubernetes.io/projected/c17ab33f-40db-4679-8e37-e3c4d013ba29-kube-api-access-jmf87\") pod \"loki-operator-controller-manager-7f7d54c9b6-mhwwb\" (UID: \"c17ab33f-40db-4679-8e37-e3c4d013ba29\") " pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.852373 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvfwj\" (UniqueName: \"kubernetes.io/projected/2778b5ad-9af6-40d9-a6a6-3e4f9c802c7d-kube-api-access-hvfwj\") pod \"cluster-logging-operator-66689c4bbf-f9skd\" (UID: \"2778b5ad-9af6-40d9-a6a6-3e4f9c802c7d\") " pod="openshift-logging/cluster-logging-operator-66689c4bbf-f9skd" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.869723 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvfwj\" (UniqueName: \"kubernetes.io/projected/2778b5ad-9af6-40d9-a6a6-3e4f9c802c7d-kube-api-access-hvfwj\") pod \"cluster-logging-operator-66689c4bbf-f9skd\" (UID: \"2778b5ad-9af6-40d9-a6a6-3e4f9c802c7d\") " pod="openshift-logging/cluster-logging-operator-66689c4bbf-f9skd" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.910760 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:25 crc kubenswrapper[4930]: I0313 09:25:25.978579 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d62b18e7-96c4-4d13-90bb-1744e16ba9ce" path="/var/lib/kubelet/pods/d62b18e7-96c4-4d13-90bb-1744e16ba9ce/volumes" Mar 13 09:25:26 crc kubenswrapper[4930]: I0313 09:25:26.064446 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-66689c4bbf-f9skd" Mar 13 09:25:26 crc kubenswrapper[4930]: I0313 09:25:26.242145 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-66689c4bbf-f9skd"] Mar 13 09:25:26 crc kubenswrapper[4930]: I0313 09:25:26.355635 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb"] Mar 13 09:25:26 crc kubenswrapper[4930]: W0313 09:25:26.360286 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc17ab33f_40db_4679_8e37_e3c4d013ba29.slice/crio-e6e94454b6de76dad2359a4de25a391ad9ae340ac0dd26368fa821635f95476a WatchSource:0}: Error finding container e6e94454b6de76dad2359a4de25a391ad9ae340ac0dd26368fa821635f95476a: Status 404 returned error can't find the container with id e6e94454b6de76dad2359a4de25a391ad9ae340ac0dd26368fa821635f95476a Mar 13 09:25:26 crc kubenswrapper[4930]: I0313 09:25:26.584276 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-66689c4bbf-f9skd" event={"ID":"2778b5ad-9af6-40d9-a6a6-3e4f9c802c7d","Type":"ContainerStarted","Data":"97eba98a5812a0b015040bd3ced10e21c3c11ce46187fc59eacb6d43492f24f8"} Mar 13 09:25:26 crc kubenswrapper[4930]: I0313 09:25:26.586487 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" event={"ID":"c17ab33f-40db-4679-8e37-e3c4d013ba29","Type":"ContainerStarted","Data":"e6e94454b6de76dad2359a4de25a391ad9ae340ac0dd26368fa821635f95476a"} Mar 13 09:25:35 crc kubenswrapper[4930]: I0313 09:25:35.671120 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-66689c4bbf-f9skd" event={"ID":"2778b5ad-9af6-40d9-a6a6-3e4f9c802c7d","Type":"ContainerStarted","Data":"cdba6f7aa704fbfda81883a54cefff0d9f0c715e01ecd434a2364c8a1ca12cc8"} Mar 13 09:25:35 crc kubenswrapper[4930]: I0313 09:25:35.672505 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" event={"ID":"c17ab33f-40db-4679-8e37-e3c4d013ba29","Type":"ContainerStarted","Data":"78f7ff6f1df0496ae8d8dd1f0636778baed2566ac4daadc0df4d18d6c78e1eed"} Mar 13 09:25:35 crc kubenswrapper[4930]: I0313 09:25:35.689741 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-66689c4bbf-f9skd" podStartSLOduration=1.8246394019999999 podStartE2EDuration="10.689723135s" podCreationTimestamp="2026-03-13 09:25:25 +0000 UTC" firstStartedPulling="2026-03-13 09:25:26.244944175 +0000 UTC m=+766.994858852" lastFinishedPulling="2026-03-13 09:25:35.110027908 +0000 UTC m=+775.859942585" observedRunningTime="2026-03-13 09:25:35.68670564 +0000 UTC m=+776.436620317" watchObservedRunningTime="2026-03-13 09:25:35.689723135 +0000 UTC m=+776.439637812" Mar 13 09:25:42 crc kubenswrapper[4930]: I0313 09:25:42.309211 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:25:42 crc kubenswrapper[4930]: I0313 09:25:42.310047 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:25:42 crc kubenswrapper[4930]: I0313 09:25:42.722874 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" event={"ID":"c17ab33f-40db-4679-8e37-e3c4d013ba29","Type":"ContainerStarted","Data":"31d5d991087d5c7dd705a5db998144015b00390dea332da6f4e24458e54c77e8"} Mar 13 09:25:42 crc kubenswrapper[4930]: I0313 09:25:42.723328 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:42 crc kubenswrapper[4930]: I0313 09:25:42.726336 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 09:25:42 crc kubenswrapper[4930]: I0313 09:25:42.759641 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" podStartSLOduration=2.544356507 podStartE2EDuration="17.759615222s" podCreationTimestamp="2026-03-13 09:25:25 +0000 UTC" firstStartedPulling="2026-03-13 09:25:26.363651893 +0000 UTC m=+767.113566560" lastFinishedPulling="2026-03-13 09:25:41.578910588 +0000 UTC m=+782.328825275" observedRunningTime="2026-03-13 09:25:42.750944287 +0000 UTC m=+783.500859024" watchObservedRunningTime="2026-03-13 09:25:42.759615222 +0000 UTC m=+783.509529909" Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.627050 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k5pqr"] Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.629143 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.633856 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k5pqr"] Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.804803 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcaa1481-0048-4477-933f-a4f11bf57c07-catalog-content\") pod \"community-operators-k5pqr\" (UID: \"fcaa1481-0048-4477-933f-a4f11bf57c07\") " pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.805026 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b485n\" (UniqueName: \"kubernetes.io/projected/fcaa1481-0048-4477-933f-a4f11bf57c07-kube-api-access-b485n\") pod \"community-operators-k5pqr\" (UID: \"fcaa1481-0048-4477-933f-a4f11bf57c07\") " pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.805137 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcaa1481-0048-4477-933f-a4f11bf57c07-utilities\") pod \"community-operators-k5pqr\" (UID: \"fcaa1481-0048-4477-933f-a4f11bf57c07\") " pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.906198 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcaa1481-0048-4477-933f-a4f11bf57c07-utilities\") pod \"community-operators-k5pqr\" (UID: \"fcaa1481-0048-4477-933f-a4f11bf57c07\") " pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.906263 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcaa1481-0048-4477-933f-a4f11bf57c07-catalog-content\") pod \"community-operators-k5pqr\" (UID: \"fcaa1481-0048-4477-933f-a4f11bf57c07\") " pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.906339 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b485n\" (UniqueName: \"kubernetes.io/projected/fcaa1481-0048-4477-933f-a4f11bf57c07-kube-api-access-b485n\") pod \"community-operators-k5pqr\" (UID: \"fcaa1481-0048-4477-933f-a4f11bf57c07\") " pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.907036 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcaa1481-0048-4477-933f-a4f11bf57c07-utilities\") pod \"community-operators-k5pqr\" (UID: \"fcaa1481-0048-4477-933f-a4f11bf57c07\") " pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.907052 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcaa1481-0048-4477-933f-a4f11bf57c07-catalog-content\") pod \"community-operators-k5pqr\" (UID: \"fcaa1481-0048-4477-933f-a4f11bf57c07\") " pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.927543 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b485n\" (UniqueName: \"kubernetes.io/projected/fcaa1481-0048-4477-933f-a4f11bf57c07-kube-api-access-b485n\") pod \"community-operators-k5pqr\" (UID: \"fcaa1481-0048-4477-933f-a4f11bf57c07\") " pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:43 crc kubenswrapper[4930]: I0313 09:25:43.945542 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:44 crc kubenswrapper[4930]: I0313 09:25:44.497860 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k5pqr"] Mar 13 09:25:44 crc kubenswrapper[4930]: I0313 09:25:44.736405 4930 generic.go:334] "Generic (PLEG): container finished" podID="fcaa1481-0048-4477-933f-a4f11bf57c07" containerID="f96a262869aac30191f312afc6a51315d060eb5bb75549061dfa30b5c7b3dc7b" exitCode=0 Mar 13 09:25:44 crc kubenswrapper[4930]: I0313 09:25:44.737898 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k5pqr" event={"ID":"fcaa1481-0048-4477-933f-a4f11bf57c07","Type":"ContainerDied","Data":"f96a262869aac30191f312afc6a51315d060eb5bb75549061dfa30b5c7b3dc7b"} Mar 13 09:25:44 crc kubenswrapper[4930]: I0313 09:25:44.738902 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k5pqr" event={"ID":"fcaa1481-0048-4477-933f-a4f11bf57c07","Type":"ContainerStarted","Data":"3d214e64830af32cd56de773aaf15925b77de6d0ceced2b997a8d8d335fe8c39"} Mar 13 09:25:45 crc kubenswrapper[4930]: I0313 09:25:45.744736 4930 generic.go:334] "Generic (PLEG): container finished" podID="fcaa1481-0048-4477-933f-a4f11bf57c07" containerID="9239ede277b808f7b022452243a7abf83f5cbddc2b042c0cb7b1e53a45b5cbd1" exitCode=0 Mar 13 09:25:45 crc kubenswrapper[4930]: I0313 09:25:45.744781 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k5pqr" event={"ID":"fcaa1481-0048-4477-933f-a4f11bf57c07","Type":"ContainerDied","Data":"9239ede277b808f7b022452243a7abf83f5cbddc2b042c0cb7b1e53a45b5cbd1"} Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.511281 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.512549 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.515816 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.516415 4930 reflector.go:368] Caches populated for *v1.Secret from object-"minio-dev"/"default-dockercfg-tkdnm" Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.516507 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.575293 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.655337 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m47hl\" (UniqueName: \"kubernetes.io/projected/ac59810a-06cb-405f-9814-7c24f9672cb7-kube-api-access-m47hl\") pod \"minio\" (UID: \"ac59810a-06cb-405f-9814-7c24f9672cb7\") " pod="minio-dev/minio" Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.655416 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-11df7e54-2f02-4560-8d90-5eb71f48e027\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11df7e54-2f02-4560-8d90-5eb71f48e027\") pod \"minio\" (UID: \"ac59810a-06cb-405f-9814-7c24f9672cb7\") " pod="minio-dev/minio" Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.755456 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k5pqr" event={"ID":"fcaa1481-0048-4477-933f-a4f11bf57c07","Type":"ContainerStarted","Data":"6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1"} Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.756837 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-11df7e54-2f02-4560-8d90-5eb71f48e027\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11df7e54-2f02-4560-8d90-5eb71f48e027\") pod \"minio\" (UID: \"ac59810a-06cb-405f-9814-7c24f9672cb7\") " pod="minio-dev/minio" Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.757023 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m47hl\" (UniqueName: \"kubernetes.io/projected/ac59810a-06cb-405f-9814-7c24f9672cb7-kube-api-access-m47hl\") pod \"minio\" (UID: \"ac59810a-06cb-405f-9814-7c24f9672cb7\") " pod="minio-dev/minio" Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.759287 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.759318 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-11df7e54-2f02-4560-8d90-5eb71f48e027\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11df7e54-2f02-4560-8d90-5eb71f48e027\") pod \"minio\" (UID: \"ac59810a-06cb-405f-9814-7c24f9672cb7\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4eae6c1d437ef73dc474c496cf7462948a6c8911e79e0ccbfc896f30f106ac16/globalmount\"" pod="minio-dev/minio" Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.787201 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m47hl\" (UniqueName: \"kubernetes.io/projected/ac59810a-06cb-405f-9814-7c24f9672cb7-kube-api-access-m47hl\") pod \"minio\" (UID: \"ac59810a-06cb-405f-9814-7c24f9672cb7\") " pod="minio-dev/minio" Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.798229 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-11df7e54-2f02-4560-8d90-5eb71f48e027\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-11df7e54-2f02-4560-8d90-5eb71f48e027\") pod \"minio\" (UID: \"ac59810a-06cb-405f-9814-7c24f9672cb7\") " pod="minio-dev/minio" Mar 13 09:25:46 crc kubenswrapper[4930]: I0313 09:25:46.863064 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Mar 13 09:25:47 crc kubenswrapper[4930]: I0313 09:25:47.250176 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k5pqr" podStartSLOduration=2.822045249 podStartE2EDuration="4.250156138s" podCreationTimestamp="2026-03-13 09:25:43 +0000 UTC" firstStartedPulling="2026-03-13 09:25:44.739603106 +0000 UTC m=+785.489517783" lastFinishedPulling="2026-03-13 09:25:46.167713985 +0000 UTC m=+786.917628672" observedRunningTime="2026-03-13 09:25:46.785418306 +0000 UTC m=+787.535332983" watchObservedRunningTime="2026-03-13 09:25:47.250156138 +0000 UTC m=+788.000070815" Mar 13 09:25:47 crc kubenswrapper[4930]: I0313 09:25:47.252518 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Mar 13 09:25:47 crc kubenswrapper[4930]: I0313 09:25:47.762827 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"ac59810a-06cb-405f-9814-7c24f9672cb7","Type":"ContainerStarted","Data":"b90295394d3b012b58b5a868ac1e9f2bdcf8109f24379ac9b6c88f8f4627981a"} Mar 13 09:25:50 crc kubenswrapper[4930]: I0313 09:25:50.784606 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"ac59810a-06cb-405f-9814-7c24f9672cb7","Type":"ContainerStarted","Data":"b23538908b55294dbf544d1f2059570a4334337c91edeec1e2ef10c01ab3cf20"} Mar 13 09:25:50 crc kubenswrapper[4930]: I0313 09:25:50.810139 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=3.550276532 podStartE2EDuration="6.810120473s" podCreationTimestamp="2026-03-13 09:25:44 +0000 UTC" firstStartedPulling="2026-03-13 09:25:47.270247277 +0000 UTC m=+788.020161954" lastFinishedPulling="2026-03-13 09:25:50.530091218 +0000 UTC m=+791.280005895" observedRunningTime="2026-03-13 09:25:50.804537184 +0000 UTC m=+791.554451891" watchObservedRunningTime="2026-03-13 09:25:50.810120473 +0000 UTC m=+791.560035170" Mar 13 09:25:53 crc kubenswrapper[4930]: I0313 09:25:53.946458 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:53 crc kubenswrapper[4930]: I0313 09:25:53.947283 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:54 crc kubenswrapper[4930]: I0313 09:25:54.008408 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:54 crc kubenswrapper[4930]: I0313 09:25:54.869141 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:54 crc kubenswrapper[4930]: I0313 09:25:54.925409 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k5pqr"] Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.678169 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-9c6b6d984-xfqch"] Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.679832 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.683206 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.683214 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.683274 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.684285 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.693519 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-rqk2d" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.698905 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-9c6b6d984-xfqch"] Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.805733 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-logging-loki-ca-bundle\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.806089 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-config\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.806133 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.806162 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-logging-loki-distributor-http\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.806228 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htsl4\" (UniqueName: \"kubernetes.io/projected/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-kube-api-access-htsl4\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.836263 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5"] Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.837157 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.839631 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.839893 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.841515 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.851871 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5"] Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.907930 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-config\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.907990 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/e5b31a3a-a3d4-4e96-9d6a-6554af218386-logging-loki-querier-http\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.908036 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.908062 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-logging-loki-distributor-http\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.908105 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/e5b31a3a-a3d4-4e96-9d6a-6554af218386-logging-loki-querier-grpc\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.908126 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5b31a3a-a3d4-4e96-9d6a-6554af218386-logging-loki-ca-bundle\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.908146 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbnv7\" (UniqueName: \"kubernetes.io/projected/e5b31a3a-a3d4-4e96-9d6a-6554af218386-kube-api-access-mbnv7\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.908170 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htsl4\" (UniqueName: \"kubernetes.io/projected/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-kube-api-access-htsl4\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.908205 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5b31a3a-a3d4-4e96-9d6a-6554af218386-config\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.908227 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/e5b31a3a-a3d4-4e96-9d6a-6554af218386-logging-loki-s3\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.908250 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-logging-loki-ca-bundle\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.908977 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-config\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.908997 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-logging-loki-ca-bundle\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.915914 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.918269 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-logging-loki-distributor-http\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.923281 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss"] Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.924056 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.929210 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.929350 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.932630 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htsl4\" (UniqueName: \"kubernetes.io/projected/e70aee0c-0c7e-428a-9fc7-c099088bd6a0-kube-api-access-htsl4\") pod \"logging-loki-distributor-9c6b6d984-xfqch\" (UID: \"e70aee0c-0c7e-428a-9fc7-c099088bd6a0\") " pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:55 crc kubenswrapper[4930]: I0313 09:25:55.947558 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss"] Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.011651 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.035304 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5b31a3a-a3d4-4e96-9d6a-6554af218386-config\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.039497 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e5b31a3a-a3d4-4e96-9d6a-6554af218386-config\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.039621 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/e5b31a3a-a3d4-4e96-9d6a-6554af218386-logging-loki-s3\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.039682 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/e5b31a3a-a3d4-4e96-9d6a-6554af218386-logging-loki-querier-http\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.039776 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/f63fa034-9ed3-4f32-9fae-c12ca9658f38-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.039807 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lvjk5\" (UniqueName: \"kubernetes.io/projected/f63fa034-9ed3-4f32-9fae-c12ca9658f38-kube-api-access-lvjk5\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.039832 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/f63fa034-9ed3-4f32-9fae-c12ca9658f38-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.039866 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f63fa034-9ed3-4f32-9fae-c12ca9658f38-config\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.039934 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/e5b31a3a-a3d4-4e96-9d6a-6554af218386-logging-loki-querier-grpc\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.039959 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5b31a3a-a3d4-4e96-9d6a-6554af218386-logging-loki-ca-bundle\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.039990 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbnv7\" (UniqueName: \"kubernetes.io/projected/e5b31a3a-a3d4-4e96-9d6a-6554af218386-kube-api-access-mbnv7\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.040042 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f63fa034-9ed3-4f32-9fae-c12ca9658f38-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.044666 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/e5b31a3a-a3d4-4e96-9d6a-6554af218386-logging-loki-querier-http\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.045512 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e5b31a3a-a3d4-4e96-9d6a-6554af218386-logging-loki-ca-bundle\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.049741 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/e5b31a3a-a3d4-4e96-9d6a-6554af218386-logging-loki-querier-grpc\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.052197 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/e5b31a3a-a3d4-4e96-9d6a-6554af218386-logging-loki-s3\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.088990 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbnv7\" (UniqueName: \"kubernetes.io/projected/e5b31a3a-a3d4-4e96-9d6a-6554af218386-kube-api-access-mbnv7\") pod \"logging-loki-querier-6dcbdf8bb8-kg8m5\" (UID: \"e5b31a3a-a3d4-4e96-9d6a-6554af218386\") " pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.099887 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh"] Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.101097 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.104722 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.105166 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.105920 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.106260 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-9ddnw" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.106514 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.106861 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.117846 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6"] Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.120215 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.141099 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh"] Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.143757 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lvjk5\" (UniqueName: \"kubernetes.io/projected/f63fa034-9ed3-4f32-9fae-c12ca9658f38-kube-api-access-lvjk5\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.143794 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/f63fa034-9ed3-4f32-9fae-c12ca9658f38-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.143836 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f63fa034-9ed3-4f32-9fae-c12ca9658f38-config\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.143914 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f63fa034-9ed3-4f32-9fae-c12ca9658f38-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.143992 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/f63fa034-9ed3-4f32-9fae-c12ca9658f38-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.150340 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f63fa034-9ed3-4f32-9fae-c12ca9658f38-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.150812 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6"] Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.151647 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f63fa034-9ed3-4f32-9fae-c12ca9658f38-config\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.156790 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/f63fa034-9ed3-4f32-9fae-c12ca9658f38-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.160607 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/f63fa034-9ed3-4f32-9fae-c12ca9658f38-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.161769 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.168740 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lvjk5\" (UniqueName: \"kubernetes.io/projected/f63fa034-9ed3-4f32-9fae-c12ca9658f38-kube-api-access-lvjk5\") pod \"logging-loki-query-frontend-ff66c4dc9-l2bss\" (UID: \"f63fa034-9ed3-4f32-9fae-c12ca9658f38\") " pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245368 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51635511-c1d7-43e1-a692-bba69266bdf7-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245419 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/bcd66746-45e7-4f01-a2ed-29da89486b3c-lokistack-gateway\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245481 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/51635511-c1d7-43e1-a692-bba69266bdf7-tls-secret\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245524 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/51635511-c1d7-43e1-a692-bba69266bdf7-tenants\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245551 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/bcd66746-45e7-4f01-a2ed-29da89486b3c-rbac\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245602 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/51635511-c1d7-43e1-a692-bba69266bdf7-lokistack-gateway\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245624 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/bcd66746-45e7-4f01-a2ed-29da89486b3c-tls-secret\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245705 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/bcd66746-45e7-4f01-a2ed-29da89486b3c-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245757 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/bcd66746-45e7-4f01-a2ed-29da89486b3c-tenants\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245820 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxdqs\" (UniqueName: \"kubernetes.io/projected/bcd66746-45e7-4f01-a2ed-29da89486b3c-kube-api-access-dxdqs\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245847 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bcd66746-45e7-4f01-a2ed-29da89486b3c-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245874 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/51635511-c1d7-43e1-a692-bba69266bdf7-rbac\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245892 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/51635511-c1d7-43e1-a692-bba69266bdf7-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.245969 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bcd66746-45e7-4f01-a2ed-29da89486b3c-logging-loki-ca-bundle\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.246010 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51635511-c1d7-43e1-a692-bba69266bdf7-logging-loki-ca-bundle\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.246028 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qt9zg\" (UniqueName: \"kubernetes.io/projected/51635511-c1d7-43e1-a692-bba69266bdf7-kube-api-access-qt9zg\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.277705 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.347472 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/bcd66746-45e7-4f01-a2ed-29da89486b3c-tls-secret\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.347516 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/bcd66746-45e7-4f01-a2ed-29da89486b3c-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.347595 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/bcd66746-45e7-4f01-a2ed-29da89486b3c-tenants\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.347629 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxdqs\" (UniqueName: \"kubernetes.io/projected/bcd66746-45e7-4f01-a2ed-29da89486b3c-kube-api-access-dxdqs\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.347732 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bcd66746-45e7-4f01-a2ed-29da89486b3c-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.347754 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/51635511-c1d7-43e1-a692-bba69266bdf7-rbac\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.347773 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/51635511-c1d7-43e1-a692-bba69266bdf7-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.348106 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bcd66746-45e7-4f01-a2ed-29da89486b3c-logging-loki-ca-bundle\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.348143 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51635511-c1d7-43e1-a692-bba69266bdf7-logging-loki-ca-bundle\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.349278 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bcd66746-45e7-4f01-a2ed-29da89486b3c-logging-loki-ca-bundle\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.349295 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/51635511-c1d7-43e1-a692-bba69266bdf7-rbac\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.349379 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bcd66746-45e7-4f01-a2ed-29da89486b3c-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.348162 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qt9zg\" (UniqueName: \"kubernetes.io/projected/51635511-c1d7-43e1-a692-bba69266bdf7-kube-api-access-qt9zg\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.349656 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51635511-c1d7-43e1-a692-bba69266bdf7-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.349705 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/bcd66746-45e7-4f01-a2ed-29da89486b3c-lokistack-gateway\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.349727 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/51635511-c1d7-43e1-a692-bba69266bdf7-tls-secret\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.349770 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/51635511-c1d7-43e1-a692-bba69266bdf7-tenants\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.349788 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/bcd66746-45e7-4f01-a2ed-29da89486b3c-rbac\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.349824 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/51635511-c1d7-43e1-a692-bba69266bdf7-lokistack-gateway\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.350681 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/bcd66746-45e7-4f01-a2ed-29da89486b3c-rbac\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.350683 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/bcd66746-45e7-4f01-a2ed-29da89486b3c-lokistack-gateway\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.352053 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/bcd66746-45e7-4f01-a2ed-29da89486b3c-tls-secret\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.352483 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/bcd66746-45e7-4f01-a2ed-29da89486b3c-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.353019 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51635511-c1d7-43e1-a692-bba69266bdf7-logging-loki-ca-bundle\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.353376 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/51635511-c1d7-43e1-a692-bba69266bdf7-lokistack-gateway\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.355022 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/51635511-c1d7-43e1-a692-bba69266bdf7-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.355711 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/51635511-c1d7-43e1-a692-bba69266bdf7-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.359762 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/bcd66746-45e7-4f01-a2ed-29da89486b3c-tenants\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.363205 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/51635511-c1d7-43e1-a692-bba69266bdf7-tls-secret\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.363566 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/51635511-c1d7-43e1-a692-bba69266bdf7-tenants\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.369380 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qt9zg\" (UniqueName: \"kubernetes.io/projected/51635511-c1d7-43e1-a692-bba69266bdf7-kube-api-access-qt9zg\") pod \"logging-loki-gateway-68bf8bdf7c-sm6c6\" (UID: \"51635511-c1d7-43e1-a692-bba69266bdf7\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.375335 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxdqs\" (UniqueName: \"kubernetes.io/projected/bcd66746-45e7-4f01-a2ed-29da89486b3c-kube-api-access-dxdqs\") pod \"logging-loki-gateway-68bf8bdf7c-zx2fh\" (UID: \"bcd66746-45e7-4f01-a2ed-29da89486b3c\") " pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.418443 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.473375 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.511261 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-9c6b6d984-xfqch"] Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.665349 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5"] Mar 13 09:25:56 crc kubenswrapper[4930]: W0313 09:25:56.674888 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5b31a3a_a3d4_4e96_9d6a_6554af218386.slice/crio-46e9fb4f62198cba8d1ff44eaf3912eedaf9444df0e1ce3a668e9b3155849de3 WatchSource:0}: Error finding container 46e9fb4f62198cba8d1ff44eaf3912eedaf9444df0e1ce3a668e9b3155849de3: Status 404 returned error can't find the container with id 46e9fb4f62198cba8d1ff44eaf3912eedaf9444df0e1ce3a668e9b3155849de3 Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.744600 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss"] Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.825010 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" event={"ID":"e70aee0c-0c7e-428a-9fc7-c099088bd6a0","Type":"ContainerStarted","Data":"330e2cc31c0b5e97cc269913d2a1932ece4aa91a5c5745fc5b2183fe34ae5ed5"} Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.827558 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" event={"ID":"f63fa034-9ed3-4f32-9fae-c12ca9658f38","Type":"ContainerStarted","Data":"784e6f46817ac2b1fa1ef9bbfff0feab3e747dc66df6259449f6178531b60ba9"} Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.828676 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" event={"ID":"e5b31a3a-a3d4-4e96-9d6a-6554af218386","Type":"ContainerStarted","Data":"46e9fb4f62198cba8d1ff44eaf3912eedaf9444df0e1ce3a668e9b3155849de3"} Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.828834 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k5pqr" podUID="fcaa1481-0048-4477-933f-a4f11bf57c07" containerName="registry-server" containerID="cri-o://6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1" gracePeriod=2 Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.844348 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.845298 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.848263 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.848533 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.851014 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.925381 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.926482 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.931777 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.932040 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Mar 13 09:25:56 crc kubenswrapper[4930]: I0313 09:25:56.965713 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.013775 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.014181 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-179db6a3-c88f-4ec8-9c3f-0d79ab7ce809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-179db6a3-c88f-4ec8-9c3f-0d79ab7ce809\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.014241 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a004ffaa-7bcb-425c-acbe-3dbd6154101d-config\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.014260 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/a004ffaa-7bcb-425c-acbe-3dbd6154101d-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.014281 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndb6m\" (UniqueName: \"kubernetes.io/projected/a004ffaa-7bcb-425c-acbe-3dbd6154101d-kube-api-access-ndb6m\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.014322 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-c01683ce-03b6-4c7c-b98e-8407cd4f5a9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c01683ce-03b6-4c7c-b98e-8407cd4f5a9a\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.014364 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/a004ffaa-7bcb-425c-acbe-3dbd6154101d-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.014405 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a004ffaa-7bcb-425c-acbe-3dbd6154101d-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.014425 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a004ffaa-7bcb-425c-acbe-3dbd6154101d-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.014655 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.018764 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.030111 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.036769 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.076941 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh"] Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.098949 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6"] Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.115908 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-c01683ce-03b6-4c7c-b98e-8407cd4f5a9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c01683ce-03b6-4c7c-b98e-8407cd4f5a9a\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116029 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/13c07815-f24e-4318-abb4-10ce48cfd113-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116100 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/a004ffaa-7bcb-425c-acbe-3dbd6154101d-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116136 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-config\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116225 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a004ffaa-7bcb-425c-acbe-3dbd6154101d-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116280 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a004ffaa-7bcb-425c-acbe-3dbd6154101d-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116301 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116356 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116446 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a004ffaa-7bcb-425c-acbe-3dbd6154101d-config\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116476 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-dd61246f-35cf-40c7-8df8-125e711819f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd61246f-35cf-40c7-8df8-125e711819f2\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116533 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116558 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c07815-f24e-4318-abb4-10ce48cfd113-config\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116613 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/13c07815-f24e-4318-abb4-10ce48cfd113-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116646 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/13c07815-f24e-4318-abb4-10ce48cfd113-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116707 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13c07815-f24e-4318-abb4-10ce48cfd113-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116730 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k6r28\" (UniqueName: \"kubernetes.io/projected/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-kube-api-access-k6r28\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116782 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116854 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-407b0d82-cfe3-4cbb-851b-783324735aa8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-407b0d82-cfe3-4cbb-851b-783324735aa8\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116922 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-179db6a3-c88f-4ec8-9c3f-0d79ab7ce809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-179db6a3-c88f-4ec8-9c3f-0d79ab7ce809\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.116954 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9jwrh\" (UniqueName: \"kubernetes.io/projected/13c07815-f24e-4318-abb4-10ce48cfd113-kube-api-access-9jwrh\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.117009 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/a004ffaa-7bcb-425c-acbe-3dbd6154101d-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.117040 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndb6m\" (UniqueName: \"kubernetes.io/projected/a004ffaa-7bcb-425c-acbe-3dbd6154101d-kube-api-access-ndb6m\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.118714 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a004ffaa-7bcb-425c-acbe-3dbd6154101d-config\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.119874 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a004ffaa-7bcb-425c-acbe-3dbd6154101d-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.122854 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/a004ffaa-7bcb-425c-acbe-3dbd6154101d-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.123372 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.123388 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/a004ffaa-7bcb-425c-acbe-3dbd6154101d-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.123399 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-179db6a3-c88f-4ec8-9c3f-0d79ab7ce809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-179db6a3-c88f-4ec8-9c3f-0d79ab7ce809\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/b34f25b1684e245b3d9d2e81a6d3a050fecf29b7b6c567b2327a22364ab9b0f5/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.123500 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.123537 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-c01683ce-03b6-4c7c-b98e-8407cd4f5a9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c01683ce-03b6-4c7c-b98e-8407cd4f5a9a\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/6491a41c8b944c6ab47bf9e93a41d5d2962d824a7e8335b123ed77228f1753a0/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.123558 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/a004ffaa-7bcb-425c-acbe-3dbd6154101d-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.134463 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndb6m\" (UniqueName: \"kubernetes.io/projected/a004ffaa-7bcb-425c-acbe-3dbd6154101d-kube-api-access-ndb6m\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.175743 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-179db6a3-c88f-4ec8-9c3f-0d79ab7ce809\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-179db6a3-c88f-4ec8-9c3f-0d79ab7ce809\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.182134 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-c01683ce-03b6-4c7c-b98e-8407cd4f5a9a\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-c01683ce-03b6-4c7c-b98e-8407cd4f5a9a\") pod \"logging-loki-ingester-0\" (UID: \"a004ffaa-7bcb-425c-acbe-3dbd6154101d\") " pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219047 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9jwrh\" (UniqueName: \"kubernetes.io/projected/13c07815-f24e-4318-abb4-10ce48cfd113-kube-api-access-9jwrh\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219112 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/13c07815-f24e-4318-abb4-10ce48cfd113-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219145 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-config\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219178 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219201 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219239 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-dd61246f-35cf-40c7-8df8-125e711819f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd61246f-35cf-40c7-8df8-125e711819f2\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219258 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219278 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c07815-f24e-4318-abb4-10ce48cfd113-config\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219302 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/13c07815-f24e-4318-abb4-10ce48cfd113-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219324 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/13c07815-f24e-4318-abb4-10ce48cfd113-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219348 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13c07815-f24e-4318-abb4-10ce48cfd113-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219367 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k6r28\" (UniqueName: \"kubernetes.io/projected/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-kube-api-access-k6r28\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219384 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.219419 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-407b0d82-cfe3-4cbb-851b-783324735aa8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-407b0d82-cfe3-4cbb-851b-783324735aa8\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.220582 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/13c07815-f24e-4318-abb4-10ce48cfd113-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.221358 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.221390 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-config\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.221876 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/13c07815-f24e-4318-abb4-10ce48cfd113-config\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.222512 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.222867 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-407b0d82-cfe3-4cbb-851b-783324735aa8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-407b0d82-cfe3-4cbb-851b-783324735aa8\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/d38fd96632125ae4eff1bff2afa164682fde1cb30e0232760ffd556f41dbe43f/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.222971 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.222996 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-dd61246f-35cf-40c7-8df8-125e711819f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd61246f-35cf-40c7-8df8-125e711819f2\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/0f485c372d6dfafae7087e84c1982bebeb20ac531e02a754ea88361f11566aa3/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.224335 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.224641 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.226617 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/13c07815-f24e-4318-abb4-10ce48cfd113-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.226828 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/13c07815-f24e-4318-abb4-10ce48cfd113-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.230084 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.231123 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/13c07815-f24e-4318-abb4-10ce48cfd113-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.236143 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9jwrh\" (UniqueName: \"kubernetes.io/projected/13c07815-f24e-4318-abb4-10ce48cfd113-kube-api-access-9jwrh\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.236986 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k6r28\" (UniqueName: \"kubernetes.io/projected/89f3b1e5-d063-4df4-93e1-5815a3ddcaf6-kube-api-access-k6r28\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.249202 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-407b0d82-cfe3-4cbb-851b-783324735aa8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-407b0d82-cfe3-4cbb-851b-783324735aa8\") pod \"logging-loki-index-gateway-0\" (UID: \"13c07815-f24e-4318-abb4-10ce48cfd113\") " pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.253976 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-dd61246f-35cf-40c7-8df8-125e711819f2\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-dd61246f-35cf-40c7-8df8-125e711819f2\") pod \"logging-loki-compactor-0\" (UID: \"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6\") " pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.294161 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.296583 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.320539 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b485n\" (UniqueName: \"kubernetes.io/projected/fcaa1481-0048-4477-933f-a4f11bf57c07-kube-api-access-b485n\") pod \"fcaa1481-0048-4477-933f-a4f11bf57c07\" (UID: \"fcaa1481-0048-4477-933f-a4f11bf57c07\") " Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.320618 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcaa1481-0048-4477-933f-a4f11bf57c07-catalog-content\") pod \"fcaa1481-0048-4477-933f-a4f11bf57c07\" (UID: \"fcaa1481-0048-4477-933f-a4f11bf57c07\") " Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.320731 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcaa1481-0048-4477-933f-a4f11bf57c07-utilities\") pod \"fcaa1481-0048-4477-933f-a4f11bf57c07\" (UID: \"fcaa1481-0048-4477-933f-a4f11bf57c07\") " Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.322662 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcaa1481-0048-4477-933f-a4f11bf57c07-utilities" (OuterVolumeSpecName: "utilities") pod "fcaa1481-0048-4477-933f-a4f11bf57c07" (UID: "fcaa1481-0048-4477-933f-a4f11bf57c07"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.325514 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fcaa1481-0048-4477-933f-a4f11bf57c07-kube-api-access-b485n" (OuterVolumeSpecName: "kube-api-access-b485n") pod "fcaa1481-0048-4477-933f-a4f11bf57c07" (UID: "fcaa1481-0048-4477-933f-a4f11bf57c07"). InnerVolumeSpecName "kube-api-access-b485n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.341040 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.384064 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fcaa1481-0048-4477-933f-a4f11bf57c07-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fcaa1481-0048-4477-933f-a4f11bf57c07" (UID: "fcaa1481-0048-4477-933f-a4f11bf57c07"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.423694 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fcaa1481-0048-4477-933f-a4f11bf57c07-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.423733 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b485n\" (UniqueName: \"kubernetes.io/projected/fcaa1481-0048-4477-933f-a4f11bf57c07-kube-api-access-b485n\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.423746 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fcaa1481-0048-4477-933f-a4f11bf57c07-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.479976 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.709858 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Mar 13 09:25:57 crc kubenswrapper[4930]: W0313 09:25:57.717538 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod89f3b1e5_d063_4df4_93e1_5815a3ddcaf6.slice/crio-5176e2386ee7b88105b831947f9b7e8b48baffa446d4fc6de879a63361ae16ad WatchSource:0}: Error finding container 5176e2386ee7b88105b831947f9b7e8b48baffa446d4fc6de879a63361ae16ad: Status 404 returned error can't find the container with id 5176e2386ee7b88105b831947f9b7e8b48baffa446d4fc6de879a63361ae16ad Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.773130 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Mar 13 09:25:57 crc kubenswrapper[4930]: W0313 09:25:57.779266 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod13c07815_f24e_4318_abb4_10ce48cfd113.slice/crio-01b93d6dcbf9e00e5b983b06c4de8f59b5e9832835822c8ac588bdb452c6ca6a WatchSource:0}: Error finding container 01b93d6dcbf9e00e5b983b06c4de8f59b5e9832835822c8ac588bdb452c6ca6a: Status 404 returned error can't find the container with id 01b93d6dcbf9e00e5b983b06c4de8f59b5e9832835822c8ac588bdb452c6ca6a Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.836720 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"13c07815-f24e-4318-abb4-10ce48cfd113","Type":"ContainerStarted","Data":"01b93d6dcbf9e00e5b983b06c4de8f59b5e9832835822c8ac588bdb452c6ca6a"} Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.838059 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" event={"ID":"51635511-c1d7-43e1-a692-bba69266bdf7","Type":"ContainerStarted","Data":"e9835cc9272f9c410c14c4a4e80adcac084d7465a67172bcfab9f3c4ba7abc2c"} Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.840159 4930 generic.go:334] "Generic (PLEG): container finished" podID="fcaa1481-0048-4477-933f-a4f11bf57c07" containerID="6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1" exitCode=0 Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.840218 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k5pqr" event={"ID":"fcaa1481-0048-4477-933f-a4f11bf57c07","Type":"ContainerDied","Data":"6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1"} Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.840240 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k5pqr" event={"ID":"fcaa1481-0048-4477-933f-a4f11bf57c07","Type":"ContainerDied","Data":"3d214e64830af32cd56de773aaf15925b77de6d0ceced2b997a8d8d335fe8c39"} Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.840259 4930 scope.go:117] "RemoveContainer" containerID="6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.840357 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k5pqr" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.844301 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6","Type":"ContainerStarted","Data":"5176e2386ee7b88105b831947f9b7e8b48baffa446d4fc6de879a63361ae16ad"} Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.845186 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" event={"ID":"bcd66746-45e7-4f01-a2ed-29da89486b3c","Type":"ContainerStarted","Data":"ec5d1631824dce31283e15ca4e861d19d65ab17dc1a3831e6f3e687b80d71fd0"} Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.874669 4930 scope.go:117] "RemoveContainer" containerID="9239ede277b808f7b022452243a7abf83f5cbddc2b042c0cb7b1e53a45b5cbd1" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.879798 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k5pqr"] Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.887475 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k5pqr"] Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.896933 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.908945 4930 scope.go:117] "RemoveContainer" containerID="f96a262869aac30191f312afc6a51315d060eb5bb75549061dfa30b5c7b3dc7b" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.979957 4930 scope.go:117] "RemoveContainer" containerID="6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1" Mar 13 09:25:57 crc kubenswrapper[4930]: E0313 09:25:57.980305 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1\": container with ID starting with 6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1 not found: ID does not exist" containerID="6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.980332 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1"} err="failed to get container status \"6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1\": rpc error: code = NotFound desc = could not find container \"6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1\": container with ID starting with 6f7b3ad9670cf4a18fd9e6aeb91940c00a0d459051b675faf7b8d9f42d8eeeb1 not found: ID does not exist" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.980354 4930 scope.go:117] "RemoveContainer" containerID="9239ede277b808f7b022452243a7abf83f5cbddc2b042c0cb7b1e53a45b5cbd1" Mar 13 09:25:57 crc kubenswrapper[4930]: E0313 09:25:57.980618 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9239ede277b808f7b022452243a7abf83f5cbddc2b042c0cb7b1e53a45b5cbd1\": container with ID starting with 9239ede277b808f7b022452243a7abf83f5cbddc2b042c0cb7b1e53a45b5cbd1 not found: ID does not exist" containerID="9239ede277b808f7b022452243a7abf83f5cbddc2b042c0cb7b1e53a45b5cbd1" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.980749 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9239ede277b808f7b022452243a7abf83f5cbddc2b042c0cb7b1e53a45b5cbd1"} err="failed to get container status \"9239ede277b808f7b022452243a7abf83f5cbddc2b042c0cb7b1e53a45b5cbd1\": rpc error: code = NotFound desc = could not find container \"9239ede277b808f7b022452243a7abf83f5cbddc2b042c0cb7b1e53a45b5cbd1\": container with ID starting with 9239ede277b808f7b022452243a7abf83f5cbddc2b042c0cb7b1e53a45b5cbd1 not found: ID does not exist" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.980847 4930 scope.go:117] "RemoveContainer" containerID="f96a262869aac30191f312afc6a51315d060eb5bb75549061dfa30b5c7b3dc7b" Mar 13 09:25:57 crc kubenswrapper[4930]: E0313 09:25:57.981181 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f96a262869aac30191f312afc6a51315d060eb5bb75549061dfa30b5c7b3dc7b\": container with ID starting with f96a262869aac30191f312afc6a51315d060eb5bb75549061dfa30b5c7b3dc7b not found: ID does not exist" containerID="f96a262869aac30191f312afc6a51315d060eb5bb75549061dfa30b5c7b3dc7b" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.981202 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f96a262869aac30191f312afc6a51315d060eb5bb75549061dfa30b5c7b3dc7b"} err="failed to get container status \"f96a262869aac30191f312afc6a51315d060eb5bb75549061dfa30b5c7b3dc7b\": rpc error: code = NotFound desc = could not find container \"f96a262869aac30191f312afc6a51315d060eb5bb75549061dfa30b5c7b3dc7b\": container with ID starting with f96a262869aac30191f312afc6a51315d060eb5bb75549061dfa30b5c7b3dc7b not found: ID does not exist" Mar 13 09:25:57 crc kubenswrapper[4930]: I0313 09:25:57.982502 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fcaa1481-0048-4477-933f-a4f11bf57c07" path="/var/lib/kubelet/pods/fcaa1481-0048-4477-933f-a4f11bf57c07/volumes" Mar 13 09:25:58 crc kubenswrapper[4930]: I0313 09:25:58.853695 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"a004ffaa-7bcb-425c-acbe-3dbd6154101d","Type":"ContainerStarted","Data":"f9dc06d10c7c212594e2787f264c80335e2f7ae917478725fa5fec13d77789b2"} Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.123723 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556566-7lxzt"] Mar 13 09:26:00 crc kubenswrapper[4930]: E0313 09:26:00.124091 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcaa1481-0048-4477-933f-a4f11bf57c07" containerName="extract-utilities" Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.124109 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcaa1481-0048-4477-933f-a4f11bf57c07" containerName="extract-utilities" Mar 13 09:26:00 crc kubenswrapper[4930]: E0313 09:26:00.124120 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcaa1481-0048-4477-933f-a4f11bf57c07" containerName="registry-server" Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.124128 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcaa1481-0048-4477-933f-a4f11bf57c07" containerName="registry-server" Mar 13 09:26:00 crc kubenswrapper[4930]: E0313 09:26:00.124147 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fcaa1481-0048-4477-933f-a4f11bf57c07" containerName="extract-content" Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.124155 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="fcaa1481-0048-4477-933f-a4f11bf57c07" containerName="extract-content" Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.124381 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="fcaa1481-0048-4477-933f-a4f11bf57c07" containerName="registry-server" Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.125008 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556566-7lxzt" Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.126674 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.128083 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.128096 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.139753 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556566-7lxzt"] Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.176372 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4dfv\" (UniqueName: \"kubernetes.io/projected/2a5d1519-2396-4ba8-983c-189a8ca117c3-kube-api-access-z4dfv\") pod \"auto-csr-approver-29556566-7lxzt\" (UID: \"2a5d1519-2396-4ba8-983c-189a8ca117c3\") " pod="openshift-infra/auto-csr-approver-29556566-7lxzt" Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.277872 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4dfv\" (UniqueName: \"kubernetes.io/projected/2a5d1519-2396-4ba8-983c-189a8ca117c3-kube-api-access-z4dfv\") pod \"auto-csr-approver-29556566-7lxzt\" (UID: \"2a5d1519-2396-4ba8-983c-189a8ca117c3\") " pod="openshift-infra/auto-csr-approver-29556566-7lxzt" Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.298993 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4dfv\" (UniqueName: \"kubernetes.io/projected/2a5d1519-2396-4ba8-983c-189a8ca117c3-kube-api-access-z4dfv\") pod \"auto-csr-approver-29556566-7lxzt\" (UID: \"2a5d1519-2396-4ba8-983c-189a8ca117c3\") " pod="openshift-infra/auto-csr-approver-29556566-7lxzt" Mar 13 09:26:00 crc kubenswrapper[4930]: I0313 09:26:00.440527 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556566-7lxzt" Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.755348 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556566-7lxzt"] Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.935079 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" event={"ID":"e70aee0c-0c7e-428a-9fc7-c099088bd6a0","Type":"ContainerStarted","Data":"021f57b01357e5875d5755ef3b3980948cb17fd38dad7153bc60a752b3108311"} Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.935185 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.937363 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"13c07815-f24e-4318-abb4-10ce48cfd113","Type":"ContainerStarted","Data":"dc8aecf2a64d1317be4375e0527558600aecd61bee60b208b0551ac93cddfff0"} Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.937582 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.938538 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556566-7lxzt" event={"ID":"2a5d1519-2396-4ba8-983c-189a8ca117c3","Type":"ContainerStarted","Data":"d4a6d572be432cd8977704ac64a1b9f7399dfc0f0787797cfa5f5ae1b7417c0d"} Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.941812 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" event={"ID":"51635511-c1d7-43e1-a692-bba69266bdf7","Type":"ContainerStarted","Data":"0fd632b83f828a4b3830f42906f7e85244c1747ebda6db053b92c44498e3597f"} Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.943101 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" event={"ID":"bcd66746-45e7-4f01-a2ed-29da89486b3c","Type":"ContainerStarted","Data":"06c83819b3e7a3432938cf117e2078229a19bde6d008850bd8c0f80e1d2fde03"} Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.944231 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" event={"ID":"f63fa034-9ed3-4f32-9fae-c12ca9658f38","Type":"ContainerStarted","Data":"31138c650c2ce637d3fa6448695f76ea113834594383e56192d332c72b357821"} Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.944376 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.945275 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"a004ffaa-7bcb-425c-acbe-3dbd6154101d","Type":"ContainerStarted","Data":"d1bb5c32c8024bf1aa40ffdb06f3ed8cd81bee40cb3b9e9eabc84751df9ca7d8"} Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.945460 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.946506 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" event={"ID":"e5b31a3a-a3d4-4e96-9d6a-6554af218386","Type":"ContainerStarted","Data":"dea875145da63a87f161d444c13fcda2a6f599bd6a4ef7474c01ec2d3dcfc4aa"} Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.946994 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.995595 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" podStartSLOduration=2.397015395 podStartE2EDuration="10.995564598s" podCreationTimestamp="2026-03-13 09:25:55 +0000 UTC" firstStartedPulling="2026-03-13 09:25:56.756565819 +0000 UTC m=+797.506480496" lastFinishedPulling="2026-03-13 09:26:05.355115022 +0000 UTC m=+806.105029699" observedRunningTime="2026-03-13 09:26:05.994233665 +0000 UTC m=+806.744148342" watchObservedRunningTime="2026-03-13 09:26:05.995564598 +0000 UTC m=+806.745479275" Mar 13 09:26:05 crc kubenswrapper[4930]: I0313 09:26:05.996715 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" podStartSLOduration=2.124701823 podStartE2EDuration="10.996708937s" podCreationTimestamp="2026-03-13 09:25:55 +0000 UTC" firstStartedPulling="2026-03-13 09:25:56.524758432 +0000 UTC m=+797.274673109" lastFinishedPulling="2026-03-13 09:26:05.396765546 +0000 UTC m=+806.146680223" observedRunningTime="2026-03-13 09:26:05.967250555 +0000 UTC m=+806.717165252" watchObservedRunningTime="2026-03-13 09:26:05.996708937 +0000 UTC m=+806.746623614" Mar 13 09:26:06 crc kubenswrapper[4930]: I0313 09:26:06.021853 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=3.479275494 podStartE2EDuration="11.021833911s" podCreationTimestamp="2026-03-13 09:25:55 +0000 UTC" firstStartedPulling="2026-03-13 09:25:57.782183861 +0000 UTC m=+798.532098538" lastFinishedPulling="2026-03-13 09:26:05.324742258 +0000 UTC m=+806.074656955" observedRunningTime="2026-03-13 09:26:06.013570975 +0000 UTC m=+806.763485652" watchObservedRunningTime="2026-03-13 09:26:06.021833911 +0000 UTC m=+806.771748588" Mar 13 09:26:06 crc kubenswrapper[4930]: I0313 09:26:06.044852 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" podStartSLOduration=2.367242346 podStartE2EDuration="11.044825732s" podCreationTimestamp="2026-03-13 09:25:55 +0000 UTC" firstStartedPulling="2026-03-13 09:25:56.681227178 +0000 UTC m=+797.431141855" lastFinishedPulling="2026-03-13 09:26:05.358810524 +0000 UTC m=+806.108725241" observedRunningTime="2026-03-13 09:26:06.043919579 +0000 UTC m=+806.793834256" watchObservedRunningTime="2026-03-13 09:26:06.044825732 +0000 UTC m=+806.794740409" Mar 13 09:26:06 crc kubenswrapper[4930]: I0313 09:26:06.970626 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"89f3b1e5-d063-4df4-93e1-5815a3ddcaf6","Type":"ContainerStarted","Data":"16c62014e2038ffa1bf6342ab982feb05b762a60c2b1fd58e82b1ed0be0410c4"} Mar 13 09:26:06 crc kubenswrapper[4930]: I0313 09:26:06.971605 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:26:07 crc kubenswrapper[4930]: I0313 09:26:07.006848 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=3.944925908 podStartE2EDuration="12.006826164s" podCreationTimestamp="2026-03-13 09:25:55 +0000 UTC" firstStartedPulling="2026-03-13 09:25:57.720922739 +0000 UTC m=+798.470837416" lastFinishedPulling="2026-03-13 09:26:05.782822995 +0000 UTC m=+806.532737672" observedRunningTime="2026-03-13 09:26:06.997231996 +0000 UTC m=+807.747146703" watchObservedRunningTime="2026-03-13 09:26:07.006826164 +0000 UTC m=+807.756740841" Mar 13 09:26:07 crc kubenswrapper[4930]: I0313 09:26:07.008027 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=4.298381207 podStartE2EDuration="12.008018784s" podCreationTimestamp="2026-03-13 09:25:55 +0000 UTC" firstStartedPulling="2026-03-13 09:25:57.914865476 +0000 UTC m=+798.664780153" lastFinishedPulling="2026-03-13 09:26:05.624503033 +0000 UTC m=+806.374417730" observedRunningTime="2026-03-13 09:26:06.06972672 +0000 UTC m=+806.819641397" watchObservedRunningTime="2026-03-13 09:26:07.008018784 +0000 UTC m=+807.757933461" Mar 13 09:26:07 crc kubenswrapper[4930]: I0313 09:26:07.979809 4930 generic.go:334] "Generic (PLEG): container finished" podID="2a5d1519-2396-4ba8-983c-189a8ca117c3" containerID="502126d38b429293d46a422c3fa582f3f7ef0f53f85780f5c35628e0822fe318" exitCode=0 Mar 13 09:26:07 crc kubenswrapper[4930]: I0313 09:26:07.980102 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556566-7lxzt" event={"ID":"2a5d1519-2396-4ba8-983c-189a8ca117c3","Type":"ContainerDied","Data":"502126d38b429293d46a422c3fa582f3f7ef0f53f85780f5c35628e0822fe318"} Mar 13 09:26:08 crc kubenswrapper[4930]: I0313 09:26:08.989987 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" event={"ID":"51635511-c1d7-43e1-a692-bba69266bdf7","Type":"ContainerStarted","Data":"8a9f024e48bf5aeddda1f1742b90e62f0ea96549a39f9acb37fccdf185484793"} Mar 13 09:26:08 crc kubenswrapper[4930]: I0313 09:26:08.990261 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:26:08 crc kubenswrapper[4930]: I0313 09:26:08.992179 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" event={"ID":"bcd66746-45e7-4f01-a2ed-29da89486b3c","Type":"ContainerStarted","Data":"b311f3a57f54236a5b697181219293ca45b0b691c6f3e1f80f7d63cc286160f9"} Mar 13 09:26:08 crc kubenswrapper[4930]: I0313 09:26:08.993031 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:26:08 crc kubenswrapper[4930]: I0313 09:26:08.993050 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:26:09 crc kubenswrapper[4930]: I0313 09:26:09.001975 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:26:09 crc kubenswrapper[4930]: I0313 09:26:09.003271 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:26:09 crc kubenswrapper[4930]: I0313 09:26:09.006047 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" Mar 13 09:26:09 crc kubenswrapper[4930]: I0313 09:26:09.017948 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podStartSLOduration=2.314712417 podStartE2EDuration="13.017926591s" podCreationTimestamp="2026-03-13 09:25:56 +0000 UTC" firstStartedPulling="2026-03-13 09:25:57.131085531 +0000 UTC m=+797.881000208" lastFinishedPulling="2026-03-13 09:26:07.834299705 +0000 UTC m=+808.584214382" observedRunningTime="2026-03-13 09:26:09.011939532 +0000 UTC m=+809.761854209" watchObservedRunningTime="2026-03-13 09:26:09.017926591 +0000 UTC m=+809.767841268" Mar 13 09:26:09 crc kubenswrapper[4930]: I0313 09:26:09.041651 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podStartSLOduration=2.3614875189999998 podStartE2EDuration="13.04162695s" podCreationTimestamp="2026-03-13 09:25:56 +0000 UTC" firstStartedPulling="2026-03-13 09:25:57.131960323 +0000 UTC m=+797.881875000" lastFinishedPulling="2026-03-13 09:26:07.812099754 +0000 UTC m=+808.562014431" observedRunningTime="2026-03-13 09:26:09.035346784 +0000 UTC m=+809.785261461" watchObservedRunningTime="2026-03-13 09:26:09.04162695 +0000 UTC m=+809.791541627" Mar 13 09:26:09 crc kubenswrapper[4930]: I0313 09:26:09.324385 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556566-7lxzt" Mar 13 09:26:09 crc kubenswrapper[4930]: I0313 09:26:09.440889 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4dfv\" (UniqueName: \"kubernetes.io/projected/2a5d1519-2396-4ba8-983c-189a8ca117c3-kube-api-access-z4dfv\") pod \"2a5d1519-2396-4ba8-983c-189a8ca117c3\" (UID: \"2a5d1519-2396-4ba8-983c-189a8ca117c3\") " Mar 13 09:26:09 crc kubenswrapper[4930]: I0313 09:26:09.447968 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a5d1519-2396-4ba8-983c-189a8ca117c3-kube-api-access-z4dfv" (OuterVolumeSpecName: "kube-api-access-z4dfv") pod "2a5d1519-2396-4ba8-983c-189a8ca117c3" (UID: "2a5d1519-2396-4ba8-983c-189a8ca117c3"). InnerVolumeSpecName "kube-api-access-z4dfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:26:09 crc kubenswrapper[4930]: I0313 09:26:09.542532 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4dfv\" (UniqueName: \"kubernetes.io/projected/2a5d1519-2396-4ba8-983c-189a8ca117c3-kube-api-access-z4dfv\") on node \"crc\" DevicePath \"\"" Mar 13 09:26:10 crc kubenswrapper[4930]: I0313 09:26:10.002645 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556566-7lxzt" event={"ID":"2a5d1519-2396-4ba8-983c-189a8ca117c3","Type":"ContainerDied","Data":"d4a6d572be432cd8977704ac64a1b9f7399dfc0f0787797cfa5f5ae1b7417c0d"} Mar 13 09:26:10 crc kubenswrapper[4930]: I0313 09:26:10.003189 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4a6d572be432cd8977704ac64a1b9f7399dfc0f0787797cfa5f5ae1b7417c0d" Mar 13 09:26:10 crc kubenswrapper[4930]: I0313 09:26:10.003220 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556566-7lxzt" Mar 13 09:26:10 crc kubenswrapper[4930]: I0313 09:26:10.003246 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:26:10 crc kubenswrapper[4930]: I0313 09:26:10.016327 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" Mar 13 09:26:10 crc kubenswrapper[4930]: I0313 09:26:10.391189 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556560-pvjlf"] Mar 13 09:26:10 crc kubenswrapper[4930]: I0313 09:26:10.405348 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556560-pvjlf"] Mar 13 09:26:11 crc kubenswrapper[4930]: I0313 09:26:11.981156 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="719d2ae6-516d-40cb-82c6-d125d2ea3f8c" path="/var/lib/kubelet/pods/719d2ae6-516d-40cb-82c6-d125d2ea3f8c/volumes" Mar 13 09:26:12 crc kubenswrapper[4930]: I0313 09:26:12.308423 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:26:12 crc kubenswrapper[4930]: I0313 09:26:12.308576 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:26:12 crc kubenswrapper[4930]: I0313 09:26:12.308678 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:26:12 crc kubenswrapper[4930]: I0313 09:26:12.310264 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e32155b41a79cf2a274b0699cb480a931ddd70ce47fbf5a1faf4047a087b2d60"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 09:26:12 crc kubenswrapper[4930]: I0313 09:26:12.310394 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://e32155b41a79cf2a274b0699cb480a931ddd70ce47fbf5a1faf4047a087b2d60" gracePeriod=600 Mar 13 09:26:13 crc kubenswrapper[4930]: I0313 09:26:13.027778 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="e32155b41a79cf2a274b0699cb480a931ddd70ce47fbf5a1faf4047a087b2d60" exitCode=0 Mar 13 09:26:13 crc kubenswrapper[4930]: I0313 09:26:13.027847 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"e32155b41a79cf2a274b0699cb480a931ddd70ce47fbf5a1faf4047a087b2d60"} Mar 13 09:26:13 crc kubenswrapper[4930]: I0313 09:26:13.028248 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"2a734be71e61d016548cb1b09f6ff35d98f17d0ce5f81f146f9d61495866a932"} Mar 13 09:26:13 crc kubenswrapper[4930]: I0313 09:26:13.028273 4930 scope.go:117] "RemoveContainer" containerID="8f2e2046f35c0bcc869ca63ddfbeac47d398ac629a37ba4258b97dd5da083a09" Mar 13 09:26:14 crc kubenswrapper[4930]: I0313 09:26:14.954772 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-764jd"] Mar 13 09:26:14 crc kubenswrapper[4930]: E0313 09:26:14.955713 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a5d1519-2396-4ba8-983c-189a8ca117c3" containerName="oc" Mar 13 09:26:14 crc kubenswrapper[4930]: I0313 09:26:14.955728 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a5d1519-2396-4ba8-983c-189a8ca117c3" containerName="oc" Mar 13 09:26:14 crc kubenswrapper[4930]: I0313 09:26:14.955941 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a5d1519-2396-4ba8-983c-189a8ca117c3" containerName="oc" Mar 13 09:26:14 crc kubenswrapper[4930]: I0313 09:26:14.957162 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:14 crc kubenswrapper[4930]: I0313 09:26:14.981948 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-764jd"] Mar 13 09:26:15 crc kubenswrapper[4930]: I0313 09:26:15.026814 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9w4v\" (UniqueName: \"kubernetes.io/projected/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-kube-api-access-p9w4v\") pod \"certified-operators-764jd\" (UID: \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\") " pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:15 crc kubenswrapper[4930]: I0313 09:26:15.027629 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-utilities\") pod \"certified-operators-764jd\" (UID: \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\") " pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:15 crc kubenswrapper[4930]: I0313 09:26:15.027699 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-catalog-content\") pod \"certified-operators-764jd\" (UID: \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\") " pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:15 crc kubenswrapper[4930]: I0313 09:26:15.129020 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9w4v\" (UniqueName: \"kubernetes.io/projected/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-kube-api-access-p9w4v\") pod \"certified-operators-764jd\" (UID: \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\") " pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:15 crc kubenswrapper[4930]: I0313 09:26:15.129159 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-utilities\") pod \"certified-operators-764jd\" (UID: \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\") " pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:15 crc kubenswrapper[4930]: I0313 09:26:15.129747 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-utilities\") pod \"certified-operators-764jd\" (UID: \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\") " pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:15 crc kubenswrapper[4930]: I0313 09:26:15.129814 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-catalog-content\") pod \"certified-operators-764jd\" (UID: \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\") " pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:15 crc kubenswrapper[4930]: I0313 09:26:15.130171 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-catalog-content\") pod \"certified-operators-764jd\" (UID: \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\") " pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:15 crc kubenswrapper[4930]: I0313 09:26:15.166656 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9w4v\" (UniqueName: \"kubernetes.io/projected/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-kube-api-access-p9w4v\") pod \"certified-operators-764jd\" (UID: \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\") " pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:15 crc kubenswrapper[4930]: I0313 09:26:15.283590 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:15 crc kubenswrapper[4930]: I0313 09:26:15.588349 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-764jd"] Mar 13 09:26:15 crc kubenswrapper[4930]: W0313 09:26:15.589467 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda26b0f53_dff7_4ad9_a8a9_e0856e710d6a.slice/crio-ccdb19bc98186354ae8d69ecfa5b5bd86a9681c754c59a09ab9fee72e5c5dbbb WatchSource:0}: Error finding container ccdb19bc98186354ae8d69ecfa5b5bd86a9681c754c59a09ab9fee72e5c5dbbb: Status 404 returned error can't find the container with id ccdb19bc98186354ae8d69ecfa5b5bd86a9681c754c59a09ab9fee72e5c5dbbb Mar 13 09:26:16 crc kubenswrapper[4930]: I0313 09:26:16.055000 4930 generic.go:334] "Generic (PLEG): container finished" podID="a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" containerID="97c84d6b7f6082fdcf81d639f53ba21d73d9be92f089b06b6f88eb07d6bc95af" exitCode=0 Mar 13 09:26:16 crc kubenswrapper[4930]: I0313 09:26:16.055057 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-764jd" event={"ID":"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a","Type":"ContainerDied","Data":"97c84d6b7f6082fdcf81d639f53ba21d73d9be92f089b06b6f88eb07d6bc95af"} Mar 13 09:26:16 crc kubenswrapper[4930]: I0313 09:26:16.055092 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-764jd" event={"ID":"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a","Type":"ContainerStarted","Data":"ccdb19bc98186354ae8d69ecfa5b5bd86a9681c754c59a09ab9fee72e5c5dbbb"} Mar 13 09:26:17 crc kubenswrapper[4930]: I0313 09:26:17.081282 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-764jd" event={"ID":"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a","Type":"ContainerStarted","Data":"98cab167b0bbf46ca1965aa9aeecf2957504e66838654bc3ce1a0e0e83de11ad"} Mar 13 09:26:18 crc kubenswrapper[4930]: I0313 09:26:18.092132 4930 generic.go:334] "Generic (PLEG): container finished" podID="a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" containerID="98cab167b0bbf46ca1965aa9aeecf2957504e66838654bc3ce1a0e0e83de11ad" exitCode=0 Mar 13 09:26:18 crc kubenswrapper[4930]: I0313 09:26:18.092248 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-764jd" event={"ID":"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a","Type":"ContainerDied","Data":"98cab167b0bbf46ca1965aa9aeecf2957504e66838654bc3ce1a0e0e83de11ad"} Mar 13 09:26:19 crc kubenswrapper[4930]: I0313 09:26:19.101341 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-764jd" event={"ID":"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a","Type":"ContainerStarted","Data":"d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd"} Mar 13 09:26:19 crc kubenswrapper[4930]: I0313 09:26:19.134983 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-764jd" podStartSLOduration=2.665923565 podStartE2EDuration="5.134962507s" podCreationTimestamp="2026-03-13 09:26:14 +0000 UTC" firstStartedPulling="2026-03-13 09:26:16.056709766 +0000 UTC m=+816.806624443" lastFinishedPulling="2026-03-13 09:26:18.525748708 +0000 UTC m=+819.275663385" observedRunningTime="2026-03-13 09:26:19.129341978 +0000 UTC m=+819.879256665" watchObservedRunningTime="2026-03-13 09:26:19.134962507 +0000 UTC m=+819.884877194" Mar 13 09:26:25 crc kubenswrapper[4930]: I0313 09:26:25.284635 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:25 crc kubenswrapper[4930]: I0313 09:26:25.285300 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:25 crc kubenswrapper[4930]: I0313 09:26:25.320692 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:26 crc kubenswrapper[4930]: I0313 09:26:26.019906 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 09:26:26 crc kubenswrapper[4930]: I0313 09:26:26.172074 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 09:26:26 crc kubenswrapper[4930]: I0313 09:26:26.211537 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:26 crc kubenswrapper[4930]: I0313 09:26:26.259179 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-764jd"] Mar 13 09:26:26 crc kubenswrapper[4930]: I0313 09:26:26.284211 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 09:26:27 crc kubenswrapper[4930]: I0313 09:26:27.301027 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Mar 13 09:26:27 crc kubenswrapper[4930]: I0313 09:26:27.352924 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 09:26:27 crc kubenswrapper[4930]: I0313 09:26:27.488211 4930 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Mar 13 09:26:27 crc kubenswrapper[4930]: I0313 09:26:27.488274 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a004ffaa-7bcb-425c-acbe-3dbd6154101d" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 13 09:26:28 crc kubenswrapper[4930]: I0313 09:26:28.175640 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-764jd" podUID="a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" containerName="registry-server" containerID="cri-o://d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd" gracePeriod=2 Mar 13 09:26:29 crc kubenswrapper[4930]: I0313 09:26:29.936670 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.084462 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-utilities\") pod \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\" (UID: \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\") " Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.084507 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-catalog-content\") pod \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\" (UID: \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\") " Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.084623 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p9w4v\" (UniqueName: \"kubernetes.io/projected/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-kube-api-access-p9w4v\") pod \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\" (UID: \"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a\") " Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.085476 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-utilities" (OuterVolumeSpecName: "utilities") pod "a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" (UID: "a26b0f53-dff7-4ad9-a8a9-e0856e710d6a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.089857 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-kube-api-access-p9w4v" (OuterVolumeSpecName: "kube-api-access-p9w4v") pod "a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" (UID: "a26b0f53-dff7-4ad9-a8a9-e0856e710d6a"). InnerVolumeSpecName "kube-api-access-p9w4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.143797 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" (UID: "a26b0f53-dff7-4ad9-a8a9-e0856e710d6a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.186602 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.186638 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.186650 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p9w4v\" (UniqueName: \"kubernetes.io/projected/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a-kube-api-access-p9w4v\") on node \"crc\" DevicePath \"\"" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.203923 4930 generic.go:334] "Generic (PLEG): container finished" podID="a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" containerID="d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd" exitCode=0 Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.203957 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-764jd" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.203967 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-764jd" event={"ID":"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a","Type":"ContainerDied","Data":"d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd"} Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.204016 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-764jd" event={"ID":"a26b0f53-dff7-4ad9-a8a9-e0856e710d6a","Type":"ContainerDied","Data":"ccdb19bc98186354ae8d69ecfa5b5bd86a9681c754c59a09ab9fee72e5c5dbbb"} Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.204039 4930 scope.go:117] "RemoveContainer" containerID="d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.219866 4930 scope.go:117] "RemoveContainer" containerID="98cab167b0bbf46ca1965aa9aeecf2957504e66838654bc3ce1a0e0e83de11ad" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.240617 4930 scope.go:117] "RemoveContainer" containerID="97c84d6b7f6082fdcf81d639f53ba21d73d9be92f089b06b6f88eb07d6bc95af" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.240944 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-764jd"] Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.252152 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-764jd"] Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.260264 4930 scope.go:117] "RemoveContainer" containerID="d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd" Mar 13 09:26:30 crc kubenswrapper[4930]: E0313 09:26:30.260656 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd\": container with ID starting with d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd not found: ID does not exist" containerID="d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.260699 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd"} err="failed to get container status \"d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd\": rpc error: code = NotFound desc = could not find container \"d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd\": container with ID starting with d3b5d836c842d40bab08d145cff41b0aae7d7ff8ef575056c1fe604faef8efdd not found: ID does not exist" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.260727 4930 scope.go:117] "RemoveContainer" containerID="98cab167b0bbf46ca1965aa9aeecf2957504e66838654bc3ce1a0e0e83de11ad" Mar 13 09:26:30 crc kubenswrapper[4930]: E0313 09:26:30.260991 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98cab167b0bbf46ca1965aa9aeecf2957504e66838654bc3ce1a0e0e83de11ad\": container with ID starting with 98cab167b0bbf46ca1965aa9aeecf2957504e66838654bc3ce1a0e0e83de11ad not found: ID does not exist" containerID="98cab167b0bbf46ca1965aa9aeecf2957504e66838654bc3ce1a0e0e83de11ad" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.261016 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98cab167b0bbf46ca1965aa9aeecf2957504e66838654bc3ce1a0e0e83de11ad"} err="failed to get container status \"98cab167b0bbf46ca1965aa9aeecf2957504e66838654bc3ce1a0e0e83de11ad\": rpc error: code = NotFound desc = could not find container \"98cab167b0bbf46ca1965aa9aeecf2957504e66838654bc3ce1a0e0e83de11ad\": container with ID starting with 98cab167b0bbf46ca1965aa9aeecf2957504e66838654bc3ce1a0e0e83de11ad not found: ID does not exist" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.261036 4930 scope.go:117] "RemoveContainer" containerID="97c84d6b7f6082fdcf81d639f53ba21d73d9be92f089b06b6f88eb07d6bc95af" Mar 13 09:26:30 crc kubenswrapper[4930]: E0313 09:26:30.261267 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97c84d6b7f6082fdcf81d639f53ba21d73d9be92f089b06b6f88eb07d6bc95af\": container with ID starting with 97c84d6b7f6082fdcf81d639f53ba21d73d9be92f089b06b6f88eb07d6bc95af not found: ID does not exist" containerID="97c84d6b7f6082fdcf81d639f53ba21d73d9be92f089b06b6f88eb07d6bc95af" Mar 13 09:26:30 crc kubenswrapper[4930]: I0313 09:26:30.261304 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97c84d6b7f6082fdcf81d639f53ba21d73d9be92f089b06b6f88eb07d6bc95af"} err="failed to get container status \"97c84d6b7f6082fdcf81d639f53ba21d73d9be92f089b06b6f88eb07d6bc95af\": rpc error: code = NotFound desc = could not find container \"97c84d6b7f6082fdcf81d639f53ba21d73d9be92f089b06b6f88eb07d6bc95af\": container with ID starting with 97c84d6b7f6082fdcf81d639f53ba21d73d9be92f089b06b6f88eb07d6bc95af not found: ID does not exist" Mar 13 09:26:31 crc kubenswrapper[4930]: I0313 09:26:31.983241 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" path="/var/lib/kubelet/pods/a26b0f53-dff7-4ad9-a8a9-e0856e710d6a/volumes" Mar 13 09:26:37 crc kubenswrapper[4930]: I0313 09:26:37.485173 4930 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Mar 13 09:26:37 crc kubenswrapper[4930]: I0313 09:26:37.485705 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a004ffaa-7bcb-425c-acbe-3dbd6154101d" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 13 09:26:41 crc kubenswrapper[4930]: I0313 09:26:41.388817 4930 scope.go:117] "RemoveContainer" containerID="d9db5a475b2e97fb11b342ee5ea95fed760a033eb6b79a64d15dc3e4af480f4a" Mar 13 09:26:47 crc kubenswrapper[4930]: I0313 09:26:47.488521 4930 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Mar 13 09:26:47 crc kubenswrapper[4930]: I0313 09:26:47.489259 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a004ffaa-7bcb-425c-acbe-3dbd6154101d" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 13 09:26:57 crc kubenswrapper[4930]: I0313 09:26:57.485523 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.238218 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-9q2fz"] Mar 13 09:27:15 crc kubenswrapper[4930]: E0313 09:27:15.239010 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" containerName="registry-server" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.239022 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" containerName="registry-server" Mar 13 09:27:15 crc kubenswrapper[4930]: E0313 09:27:15.239033 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" containerName="extract-content" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.239039 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" containerName="extract-content" Mar 13 09:27:15 crc kubenswrapper[4930]: E0313 09:27:15.239055 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" containerName="extract-utilities" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.239063 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" containerName="extract-utilities" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.239192 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="a26b0f53-dff7-4ad9-a8a9-e0856e710d6a" containerName="registry-server" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.239680 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.244760 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.245346 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.245568 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.246020 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-wjgqb" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.247185 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.253532 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.265944 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-9q2fz"] Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.350768 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-metrics\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.350839 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-trusted-ca\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.350884 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-syslog-receiver\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.350910 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-token\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.350948 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-config-openshift-service-cacrt\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.351054 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-sa-token\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.351085 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-tmp\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.351123 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-entrypoint\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.351151 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-config\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.351171 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlv8j\" (UniqueName: \"kubernetes.io/projected/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-kube-api-access-tlv8j\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.351214 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-datadir\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.392716 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-9q2fz"] Mar 13 09:27:15 crc kubenswrapper[4930]: E0313 09:27:15.393310 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-tlv8j metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-9q2fz" podUID="19aa5676-f99f-4cd5-9c2d-4cb56db33f27" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.452558 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-config-openshift-service-cacrt\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.452622 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-sa-token\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.452649 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-tmp\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.452679 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-entrypoint\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.452705 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-config\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.452727 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlv8j\" (UniqueName: \"kubernetes.io/projected/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-kube-api-access-tlv8j\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.452773 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-datadir\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.452804 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-metrics\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.452829 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-trusted-ca\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.452858 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-syslog-receiver\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.452885 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-token\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: E0313 09:27:15.453242 4930 secret.go:188] Couldn't get secret openshift-logging/collector-metrics: secret "collector-metrics" not found Mar 13 09:27:15 crc kubenswrapper[4930]: E0313 09:27:15.453316 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-metrics podName:19aa5676-f99f-4cd5-9c2d-4cb56db33f27 nodeName:}" failed. No retries permitted until 2026-03-13 09:27:15.953295839 +0000 UTC m=+876.703210536 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics" (UniqueName: "kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-metrics") pod "collector-9q2fz" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27") : secret "collector-metrics" not found Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.453461 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-datadir\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: E0313 09:27:15.453543 4930 secret.go:188] Couldn't get secret openshift-logging/collector-syslog-receiver: secret "collector-syslog-receiver" not found Mar 13 09:27:15 crc kubenswrapper[4930]: E0313 09:27:15.453634 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-syslog-receiver podName:19aa5676-f99f-4cd5-9c2d-4cb56db33f27 nodeName:}" failed. No retries permitted until 2026-03-13 09:27:15.953614497 +0000 UTC m=+876.703529174 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "collector-syslog-receiver" (UniqueName: "kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-syslog-receiver") pod "collector-9q2fz" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27") : secret "collector-syslog-receiver" not found Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.454399 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-config-openshift-service-cacrt\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.454534 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-entrypoint\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.454877 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-trusted-ca\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.454941 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-config\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.460681 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-token\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.469849 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-tmp\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.472392 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlv8j\" (UniqueName: \"kubernetes.io/projected/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-kube-api-access-tlv8j\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.473780 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-sa-token\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.573955 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.583015 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.654934 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-datadir\") pod \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.655227 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-config\") pod \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.655383 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-token\") pod \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.655538 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-config-openshift-service-cacrt\") pod \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.655661 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-trusted-ca\") pod \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.655801 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-sa-token\") pod \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.656192 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-entrypoint\") pod \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.656312 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlv8j\" (UniqueName: \"kubernetes.io/projected/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-kube-api-access-tlv8j\") pod \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.656450 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-tmp\") pod \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.655058 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-datadir" (OuterVolumeSpecName: "datadir") pod "19aa5676-f99f-4cd5-9c2d-4cb56db33f27" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.656582 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "19aa5676-f99f-4cd5-9c2d-4cb56db33f27" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.655757 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-config" (OuterVolumeSpecName: "config") pod "19aa5676-f99f-4cd5-9c2d-4cb56db33f27" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.655974 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "19aa5676-f99f-4cd5-9c2d-4cb56db33f27" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.656060 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "19aa5676-f99f-4cd5-9c2d-4cb56db33f27" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.657193 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.657321 4930 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-entrypoint\") on node \"crc\" DevicePath \"\"" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.657397 4930 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-datadir\") on node \"crc\" DevicePath \"\"" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.657493 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.657609 4930 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.659009 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-tmp" (OuterVolumeSpecName: "tmp") pod "19aa5676-f99f-4cd5-9c2d-4cb56db33f27" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.659363 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-kube-api-access-tlv8j" (OuterVolumeSpecName: "kube-api-access-tlv8j") pod "19aa5676-f99f-4cd5-9c2d-4cb56db33f27" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27"). InnerVolumeSpecName "kube-api-access-tlv8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.659462 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-token" (OuterVolumeSpecName: "collector-token") pod "19aa5676-f99f-4cd5-9c2d-4cb56db33f27" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.662655 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-sa-token" (OuterVolumeSpecName: "sa-token") pod "19aa5676-f99f-4cd5-9c2d-4cb56db33f27" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.759043 4930 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-token\") on node \"crc\" DevicePath \"\"" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.759071 4930 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-sa-token\") on node \"crc\" DevicePath \"\"" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.759079 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlv8j\" (UniqueName: \"kubernetes.io/projected/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-kube-api-access-tlv8j\") on node \"crc\" DevicePath \"\"" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.759087 4930 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-tmp\") on node \"crc\" DevicePath \"\"" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.962179 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-metrics\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.962248 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-syslog-receiver\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.964889 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-metrics\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:15 crc kubenswrapper[4930]: I0313 09:27:15.965003 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-syslog-receiver\") pod \"collector-9q2fz\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " pod="openshift-logging/collector-9q2fz" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.063062 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-syslog-receiver\") pod \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.063121 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-metrics\") pod \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\" (UID: \"19aa5676-f99f-4cd5-9c2d-4cb56db33f27\") " Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.066051 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-metrics" (OuterVolumeSpecName: "metrics") pod "19aa5676-f99f-4cd5-9c2d-4cb56db33f27" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.066190 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "19aa5676-f99f-4cd5-9c2d-4cb56db33f27" (UID: "19aa5676-f99f-4cd5-9c2d-4cb56db33f27"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.164546 4930 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.164583 4930 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/19aa5676-f99f-4cd5-9c2d-4cb56db33f27-metrics\") on node \"crc\" DevicePath \"\"" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.580405 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-9q2fz" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.640479 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-9q2fz"] Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.649359 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-9q2fz"] Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.655334 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-dz2f6"] Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.656393 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.663626 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-dz2f6"] Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.663870 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.664074 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.664177 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.664566 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-wjgqb" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.666307 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.668674 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.771634 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f94fba37-ce39-45bc-b772-a0efb6b39d63-tmp\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.772235 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/f94fba37-ce39-45bc-b772-a0efb6b39d63-collector-syslog-receiver\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.772356 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f94fba37-ce39-45bc-b772-a0efb6b39d63-trusted-ca\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.772484 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/f94fba37-ce39-45bc-b772-a0efb6b39d63-sa-token\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.772606 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/f94fba37-ce39-45bc-b772-a0efb6b39d63-metrics\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.772716 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dv77g\" (UniqueName: \"kubernetes.io/projected/f94fba37-ce39-45bc-b772-a0efb6b39d63-kube-api-access-dv77g\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.772843 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/f94fba37-ce39-45bc-b772-a0efb6b39d63-datadir\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.772965 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/f94fba37-ce39-45bc-b772-a0efb6b39d63-entrypoint\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.773093 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/f94fba37-ce39-45bc-b772-a0efb6b39d63-config-openshift-service-cacrt\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.773206 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f94fba37-ce39-45bc-b772-a0efb6b39d63-config\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.773312 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/f94fba37-ce39-45bc-b772-a0efb6b39d63-collector-token\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.874811 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dv77g\" (UniqueName: \"kubernetes.io/projected/f94fba37-ce39-45bc-b772-a0efb6b39d63-kube-api-access-dv77g\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.875135 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/f94fba37-ce39-45bc-b772-a0efb6b39d63-datadir\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.875276 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/f94fba37-ce39-45bc-b772-a0efb6b39d63-entrypoint\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.875388 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/f94fba37-ce39-45bc-b772-a0efb6b39d63-config-openshift-service-cacrt\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.875545 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f94fba37-ce39-45bc-b772-a0efb6b39d63-config\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.875670 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/f94fba37-ce39-45bc-b772-a0efb6b39d63-collector-token\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.875788 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f94fba37-ce39-45bc-b772-a0efb6b39d63-tmp\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.875915 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/f94fba37-ce39-45bc-b772-a0efb6b39d63-collector-syslog-receiver\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.876021 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f94fba37-ce39-45bc-b772-a0efb6b39d63-trusted-ca\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.876130 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/f94fba37-ce39-45bc-b772-a0efb6b39d63-sa-token\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.876214 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/f94fba37-ce39-45bc-b772-a0efb6b39d63-config-openshift-service-cacrt\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.875208 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/f94fba37-ce39-45bc-b772-a0efb6b39d63-datadir\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.876295 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/f94fba37-ce39-45bc-b772-a0efb6b39d63-metrics\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.876414 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f94fba37-ce39-45bc-b772-a0efb6b39d63-config\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.877084 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f94fba37-ce39-45bc-b772-a0efb6b39d63-trusted-ca\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.877332 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/f94fba37-ce39-45bc-b772-a0efb6b39d63-entrypoint\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.878945 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/f94fba37-ce39-45bc-b772-a0efb6b39d63-tmp\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.880018 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/f94fba37-ce39-45bc-b772-a0efb6b39d63-collector-syslog-receiver\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.880988 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/f94fba37-ce39-45bc-b772-a0efb6b39d63-collector-token\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.890564 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dv77g\" (UniqueName: \"kubernetes.io/projected/f94fba37-ce39-45bc-b772-a0efb6b39d63-kube-api-access-dv77g\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.890755 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/f94fba37-ce39-45bc-b772-a0efb6b39d63-metrics\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.896637 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/f94fba37-ce39-45bc-b772-a0efb6b39d63-sa-token\") pod \"collector-dz2f6\" (UID: \"f94fba37-ce39-45bc-b772-a0efb6b39d63\") " pod="openshift-logging/collector-dz2f6" Mar 13 09:27:16 crc kubenswrapper[4930]: I0313 09:27:16.976765 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-dz2f6" Mar 13 09:27:17 crc kubenswrapper[4930]: I0313 09:27:17.384981 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-dz2f6"] Mar 13 09:27:17 crc kubenswrapper[4930]: I0313 09:27:17.587381 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-dz2f6" event={"ID":"f94fba37-ce39-45bc-b772-a0efb6b39d63","Type":"ContainerStarted","Data":"61c1a44ef28a72b7cf9df08a746282af4c5b61a7e83dfb10ca5359b83c8a94ad"} Mar 13 09:27:17 crc kubenswrapper[4930]: I0313 09:27:17.981856 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19aa5676-f99f-4cd5-9c2d-4cb56db33f27" path="/var/lib/kubelet/pods/19aa5676-f99f-4cd5-9c2d-4cb56db33f27/volumes" Mar 13 09:27:21 crc kubenswrapper[4930]: I0313 09:27:21.616832 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-dz2f6" event={"ID":"f94fba37-ce39-45bc-b772-a0efb6b39d63","Type":"ContainerStarted","Data":"c578475ff12db639e4136ae2e659dca29cf5464683312da6a7d1186f036f277b"} Mar 13 09:27:21 crc kubenswrapper[4930]: I0313 09:27:21.636080 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-dz2f6" podStartSLOduration=1.992086682 podStartE2EDuration="5.636062373s" podCreationTimestamp="2026-03-13 09:27:16 +0000 UTC" firstStartedPulling="2026-03-13 09:27:17.39084777 +0000 UTC m=+878.140762447" lastFinishedPulling="2026-03-13 09:27:21.034823451 +0000 UTC m=+881.784738138" observedRunningTime="2026-03-13 09:27:21.633041898 +0000 UTC m=+882.382956575" watchObservedRunningTime="2026-03-13 09:27:21.636062373 +0000 UTC m=+882.385977050" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.168012 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl"] Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.171599 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.174130 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.198613 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl"] Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.283535 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8x7vg\" (UniqueName: \"kubernetes.io/projected/82b73078-bbcf-4068-879a-04602eae8c47-kube-api-access-8x7vg\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl\" (UID: \"82b73078-bbcf-4068-879a-04602eae8c47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.283592 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82b73078-bbcf-4068-879a-04602eae8c47-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl\" (UID: \"82b73078-bbcf-4068-879a-04602eae8c47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.283646 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82b73078-bbcf-4068-879a-04602eae8c47-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl\" (UID: \"82b73078-bbcf-4068-879a-04602eae8c47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.385790 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8x7vg\" (UniqueName: \"kubernetes.io/projected/82b73078-bbcf-4068-879a-04602eae8c47-kube-api-access-8x7vg\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl\" (UID: \"82b73078-bbcf-4068-879a-04602eae8c47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.385849 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82b73078-bbcf-4068-879a-04602eae8c47-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl\" (UID: \"82b73078-bbcf-4068-879a-04602eae8c47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.385903 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82b73078-bbcf-4068-879a-04602eae8c47-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl\" (UID: \"82b73078-bbcf-4068-879a-04602eae8c47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.386347 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82b73078-bbcf-4068-879a-04602eae8c47-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl\" (UID: \"82b73078-bbcf-4068-879a-04602eae8c47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.386839 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82b73078-bbcf-4068-879a-04602eae8c47-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl\" (UID: \"82b73078-bbcf-4068-879a-04602eae8c47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.404287 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8x7vg\" (UniqueName: \"kubernetes.io/projected/82b73078-bbcf-4068-879a-04602eae8c47-kube-api-access-8x7vg\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl\" (UID: \"82b73078-bbcf-4068-879a-04602eae8c47\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.500002 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:27:54 crc kubenswrapper[4930]: I0313 09:27:54.972061 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl"] Mar 13 09:27:55 crc kubenswrapper[4930]: I0313 09:27:55.890153 4930 generic.go:334] "Generic (PLEG): container finished" podID="82b73078-bbcf-4068-879a-04602eae8c47" containerID="da9fbdc8d4e4036d66ac35d4754262cea79eb5f14b59fbf11664a3c3324b4307" exitCode=0 Mar 13 09:27:55 crc kubenswrapper[4930]: I0313 09:27:55.890217 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" event={"ID":"82b73078-bbcf-4068-879a-04602eae8c47","Type":"ContainerDied","Data":"da9fbdc8d4e4036d66ac35d4754262cea79eb5f14b59fbf11664a3c3324b4307"} Mar 13 09:27:55 crc kubenswrapper[4930]: I0313 09:27:55.890409 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" event={"ID":"82b73078-bbcf-4068-879a-04602eae8c47","Type":"ContainerStarted","Data":"32aff494040e6bbe1f3c8f5064ccc05454f311facf66fa047d371bcba2dd8a7f"} Mar 13 09:27:57 crc kubenswrapper[4930]: I0313 09:27:57.906745 4930 generic.go:334] "Generic (PLEG): container finished" podID="82b73078-bbcf-4068-879a-04602eae8c47" containerID="38b38b489b08d193da67724b4f314486992af6758d791db1ce8db3cb2f91ea81" exitCode=0 Mar 13 09:27:57 crc kubenswrapper[4930]: I0313 09:27:57.906884 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" event={"ID":"82b73078-bbcf-4068-879a-04602eae8c47","Type":"ContainerDied","Data":"38b38b489b08d193da67724b4f314486992af6758d791db1ce8db3cb2f91ea81"} Mar 13 09:27:58 crc kubenswrapper[4930]: I0313 09:27:58.918390 4930 generic.go:334] "Generic (PLEG): container finished" podID="82b73078-bbcf-4068-879a-04602eae8c47" containerID="a559a5204a8d90f660e6106f426af88c7ea506055c52ecc3cf9560da51a54eda" exitCode=0 Mar 13 09:27:58 crc kubenswrapper[4930]: I0313 09:27:58.918510 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" event={"ID":"82b73078-bbcf-4068-879a-04602eae8c47","Type":"ContainerDied","Data":"a559a5204a8d90f660e6106f426af88c7ea506055c52ecc3cf9560da51a54eda"} Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.133470 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556568-dvxwm"] Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.134957 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556568-dvxwm" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.138993 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.147793 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.147804 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.156448 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556568-dvxwm"] Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.271369 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx5m6\" (UniqueName: \"kubernetes.io/projected/8210f60a-65d8-4fa0-9ad7-c187c3393c5e-kube-api-access-mx5m6\") pod \"auto-csr-approver-29556568-dvxwm\" (UID: \"8210f60a-65d8-4fa0-9ad7-c187c3393c5e\") " pod="openshift-infra/auto-csr-approver-29556568-dvxwm" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.324229 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.373052 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx5m6\" (UniqueName: \"kubernetes.io/projected/8210f60a-65d8-4fa0-9ad7-c187c3393c5e-kube-api-access-mx5m6\") pod \"auto-csr-approver-29556568-dvxwm\" (UID: \"8210f60a-65d8-4fa0-9ad7-c187c3393c5e\") " pod="openshift-infra/auto-csr-approver-29556568-dvxwm" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.390879 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx5m6\" (UniqueName: \"kubernetes.io/projected/8210f60a-65d8-4fa0-9ad7-c187c3393c5e-kube-api-access-mx5m6\") pod \"auto-csr-approver-29556568-dvxwm\" (UID: \"8210f60a-65d8-4fa0-9ad7-c187c3393c5e\") " pod="openshift-infra/auto-csr-approver-29556568-dvxwm" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.458372 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556568-dvxwm" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.474273 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82b73078-bbcf-4068-879a-04602eae8c47-bundle\") pod \"82b73078-bbcf-4068-879a-04602eae8c47\" (UID: \"82b73078-bbcf-4068-879a-04602eae8c47\") " Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.474379 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82b73078-bbcf-4068-879a-04602eae8c47-util\") pod \"82b73078-bbcf-4068-879a-04602eae8c47\" (UID: \"82b73078-bbcf-4068-879a-04602eae8c47\") " Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.474420 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8x7vg\" (UniqueName: \"kubernetes.io/projected/82b73078-bbcf-4068-879a-04602eae8c47-kube-api-access-8x7vg\") pod \"82b73078-bbcf-4068-879a-04602eae8c47\" (UID: \"82b73078-bbcf-4068-879a-04602eae8c47\") " Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.474898 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82b73078-bbcf-4068-879a-04602eae8c47-bundle" (OuterVolumeSpecName: "bundle") pod "82b73078-bbcf-4068-879a-04602eae8c47" (UID: "82b73078-bbcf-4068-879a-04602eae8c47"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.477679 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82b73078-bbcf-4068-879a-04602eae8c47-kube-api-access-8x7vg" (OuterVolumeSpecName: "kube-api-access-8x7vg") pod "82b73078-bbcf-4068-879a-04602eae8c47" (UID: "82b73078-bbcf-4068-879a-04602eae8c47"). InnerVolumeSpecName "kube-api-access-8x7vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.576149 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8x7vg\" (UniqueName: \"kubernetes.io/projected/82b73078-bbcf-4068-879a-04602eae8c47-kube-api-access-8x7vg\") on node \"crc\" DevicePath \"\"" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.576189 4930 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/82b73078-bbcf-4068-879a-04602eae8c47-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.770447 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82b73078-bbcf-4068-879a-04602eae8c47-util" (OuterVolumeSpecName: "util") pod "82b73078-bbcf-4068-879a-04602eae8c47" (UID: "82b73078-bbcf-4068-879a-04602eae8c47"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.785839 4930 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/82b73078-bbcf-4068-879a-04602eae8c47-util\") on node \"crc\" DevicePath \"\"" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.860941 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556568-dvxwm"] Mar 13 09:28:00 crc kubenswrapper[4930]: W0313 09:28:00.865111 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8210f60a_65d8_4fa0_9ad7_c187c3393c5e.slice/crio-f32429787a1d4822b5e90f687418c50eaefe504d41bcc2264e7ab9faae457eb7 WatchSource:0}: Error finding container f32429787a1d4822b5e90f687418c50eaefe504d41bcc2264e7ab9faae457eb7: Status 404 returned error can't find the container with id f32429787a1d4822b5e90f687418c50eaefe504d41bcc2264e7ab9faae457eb7 Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.934522 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.934511 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jm7tl" event={"ID":"82b73078-bbcf-4068-879a-04602eae8c47","Type":"ContainerDied","Data":"32aff494040e6bbe1f3c8f5064ccc05454f311facf66fa047d371bcba2dd8a7f"} Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.934633 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32aff494040e6bbe1f3c8f5064ccc05454f311facf66fa047d371bcba2dd8a7f" Mar 13 09:28:00 crc kubenswrapper[4930]: I0313 09:28:00.935877 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556568-dvxwm" event={"ID":"8210f60a-65d8-4fa0-9ad7-c187c3393c5e","Type":"ContainerStarted","Data":"f32429787a1d4822b5e90f687418c50eaefe504d41bcc2264e7ab9faae457eb7"} Mar 13 09:28:01 crc kubenswrapper[4930]: I0313 09:28:01.943620 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556568-dvxwm" event={"ID":"8210f60a-65d8-4fa0-9ad7-c187c3393c5e","Type":"ContainerStarted","Data":"d9999ba5aeb068d6c679d90f73c0151d561e7563d00ab3a7c4d7da1e740dc633"} Mar 13 09:28:01 crc kubenswrapper[4930]: I0313 09:28:01.955342 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556568-dvxwm" podStartSLOduration=1.179138484 podStartE2EDuration="1.955324612s" podCreationTimestamp="2026-03-13 09:28:00 +0000 UTC" firstStartedPulling="2026-03-13 09:28:00.868624212 +0000 UTC m=+921.618538899" lastFinishedPulling="2026-03-13 09:28:01.64481035 +0000 UTC m=+922.394725027" observedRunningTime="2026-03-13 09:28:01.954355708 +0000 UTC m=+922.704270385" watchObservedRunningTime="2026-03-13 09:28:01.955324612 +0000 UTC m=+922.705239289" Mar 13 09:28:02 crc kubenswrapper[4930]: I0313 09:28:02.953159 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556568-dvxwm" event={"ID":"8210f60a-65d8-4fa0-9ad7-c187c3393c5e","Type":"ContainerDied","Data":"d9999ba5aeb068d6c679d90f73c0151d561e7563d00ab3a7c4d7da1e740dc633"} Mar 13 09:28:02 crc kubenswrapper[4930]: I0313 09:28:02.953009 4930 generic.go:334] "Generic (PLEG): container finished" podID="8210f60a-65d8-4fa0-9ad7-c187c3393c5e" containerID="d9999ba5aeb068d6c679d90f73c0151d561e7563d00ab3a7c4d7da1e740dc633" exitCode=0 Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.205407 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-h7td2"] Mar 13 09:28:03 crc kubenswrapper[4930]: E0313 09:28:03.211134 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b73078-bbcf-4068-879a-04602eae8c47" containerName="extract" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.211172 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b73078-bbcf-4068-879a-04602eae8c47" containerName="extract" Mar 13 09:28:03 crc kubenswrapper[4930]: E0313 09:28:03.211211 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b73078-bbcf-4068-879a-04602eae8c47" containerName="util" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.211220 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b73078-bbcf-4068-879a-04602eae8c47" containerName="util" Mar 13 09:28:03 crc kubenswrapper[4930]: E0313 09:28:03.211261 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82b73078-bbcf-4068-879a-04602eae8c47" containerName="pull" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.211270 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="82b73078-bbcf-4068-879a-04602eae8c47" containerName="pull" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.211679 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="82b73078-bbcf-4068-879a-04602eae8c47" containerName="extract" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.213194 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-h7td2" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.220212 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.220276 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.220567 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-ckxb6" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.234894 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-h7td2"] Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.323044 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2rm7b\" (UniqueName: \"kubernetes.io/projected/33e7ce13-4b1a-4274-83f6-93e2c27e1abb-kube-api-access-2rm7b\") pod \"nmstate-operator-796d4cfff4-h7td2\" (UID: \"33e7ce13-4b1a-4274-83f6-93e2c27e1abb\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-h7td2" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.425055 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2rm7b\" (UniqueName: \"kubernetes.io/projected/33e7ce13-4b1a-4274-83f6-93e2c27e1abb-kube-api-access-2rm7b\") pod \"nmstate-operator-796d4cfff4-h7td2\" (UID: \"33e7ce13-4b1a-4274-83f6-93e2c27e1abb\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-h7td2" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.441759 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2rm7b\" (UniqueName: \"kubernetes.io/projected/33e7ce13-4b1a-4274-83f6-93e2c27e1abb-kube-api-access-2rm7b\") pod \"nmstate-operator-796d4cfff4-h7td2\" (UID: \"33e7ce13-4b1a-4274-83f6-93e2c27e1abb\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-h7td2" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.538954 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-h7td2" Mar 13 09:28:03 crc kubenswrapper[4930]: I0313 09:28:03.994085 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-h7td2"] Mar 13 09:28:04 crc kubenswrapper[4930]: I0313 09:28:04.230623 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556568-dvxwm" Mar 13 09:28:04 crc kubenswrapper[4930]: I0313 09:28:04.345869 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mx5m6\" (UniqueName: \"kubernetes.io/projected/8210f60a-65d8-4fa0-9ad7-c187c3393c5e-kube-api-access-mx5m6\") pod \"8210f60a-65d8-4fa0-9ad7-c187c3393c5e\" (UID: \"8210f60a-65d8-4fa0-9ad7-c187c3393c5e\") " Mar 13 09:28:04 crc kubenswrapper[4930]: I0313 09:28:04.350700 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8210f60a-65d8-4fa0-9ad7-c187c3393c5e-kube-api-access-mx5m6" (OuterVolumeSpecName: "kube-api-access-mx5m6") pod "8210f60a-65d8-4fa0-9ad7-c187c3393c5e" (UID: "8210f60a-65d8-4fa0-9ad7-c187c3393c5e"). InnerVolumeSpecName "kube-api-access-mx5m6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:28:04 crc kubenswrapper[4930]: I0313 09:28:04.448384 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mx5m6\" (UniqueName: \"kubernetes.io/projected/8210f60a-65d8-4fa0-9ad7-c187c3393c5e-kube-api-access-mx5m6\") on node \"crc\" DevicePath \"\"" Mar 13 09:28:04 crc kubenswrapper[4930]: I0313 09:28:04.968512 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556568-dvxwm" event={"ID":"8210f60a-65d8-4fa0-9ad7-c187c3393c5e","Type":"ContainerDied","Data":"f32429787a1d4822b5e90f687418c50eaefe504d41bcc2264e7ab9faae457eb7"} Mar 13 09:28:04 crc kubenswrapper[4930]: I0313 09:28:04.968797 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f32429787a1d4822b5e90f687418c50eaefe504d41bcc2264e7ab9faae457eb7" Mar 13 09:28:04 crc kubenswrapper[4930]: I0313 09:28:04.968930 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556568-dvxwm" Mar 13 09:28:04 crc kubenswrapper[4930]: I0313 09:28:04.980478 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-h7td2" event={"ID":"33e7ce13-4b1a-4274-83f6-93e2c27e1abb","Type":"ContainerStarted","Data":"dcb06dc5f9ddedf94e11dc4718f32b1e96b3a4890cca73ede35a636d66f02347"} Mar 13 09:28:05 crc kubenswrapper[4930]: I0313 09:28:05.042718 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556562-k9kvz"] Mar 13 09:28:05 crc kubenswrapper[4930]: I0313 09:28:05.047623 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556562-k9kvz"] Mar 13 09:28:05 crc kubenswrapper[4930]: I0313 09:28:05.980258 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="347caee5-9072-4e3a-946c-09dcf974d068" path="/var/lib/kubelet/pods/347caee5-9072-4e3a-946c-09dcf974d068/volumes" Mar 13 09:28:06 crc kubenswrapper[4930]: I0313 09:28:06.997825 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-h7td2" event={"ID":"33e7ce13-4b1a-4274-83f6-93e2c27e1abb","Type":"ContainerStarted","Data":"855298a16c35e442a9bb46b9b54b7e6a42c51848708b0fb82426179aebfd8817"} Mar 13 09:28:07 crc kubenswrapper[4930]: I0313 09:28:07.025573 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-h7td2" podStartSLOduration=1.537598085 podStartE2EDuration="4.025544465s" podCreationTimestamp="2026-03-13 09:28:03 +0000 UTC" firstStartedPulling="2026-03-13 09:28:03.990333043 +0000 UTC m=+924.740247720" lastFinishedPulling="2026-03-13 09:28:06.478279423 +0000 UTC m=+927.228194100" observedRunningTime="2026-03-13 09:28:07.014427739 +0000 UTC m=+927.764342416" watchObservedRunningTime="2026-03-13 09:28:07.025544465 +0000 UTC m=+927.775459162" Mar 13 09:28:12 crc kubenswrapper[4930]: I0313 09:28:12.308611 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:28:12 crc kubenswrapper[4930]: I0313 09:28:12.308883 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.860213 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-z6585"] Mar 13 09:28:13 crc kubenswrapper[4930]: E0313 09:28:13.860812 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8210f60a-65d8-4fa0-9ad7-c187c3393c5e" containerName="oc" Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.860827 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="8210f60a-65d8-4fa0-9ad7-c187c3393c5e" containerName="oc" Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.860956 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="8210f60a-65d8-4fa0-9ad7-c187c3393c5e" containerName="oc" Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.861408 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.865960 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-sr796" Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.865974 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.870099 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-h8l9q"] Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.871587 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-h8l9q" Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.875867 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-z6585"] Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.882703 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-h8l9q"] Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.896133 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nsvm\" (UniqueName: \"kubernetes.io/projected/ca17bd84-571a-44a2-b882-1372cdc91ab2-kube-api-access-2nsvm\") pod \"nmstate-metrics-9b8c8685d-h8l9q\" (UID: \"ca17bd84-571a-44a2-b882-1372cdc91ab2\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-h8l9q" Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.908998 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-zrwjj"] Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.909870 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.997373 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9fc4b697-3810-4744-8c3f-2285ee8399e8-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-z6585\" (UID: \"9fc4b697-3810-4744-8c3f-2285ee8399e8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.997426 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bf8zh\" (UniqueName: \"kubernetes.io/projected/9fc4b697-3810-4744-8c3f-2285ee8399e8-kube-api-access-bf8zh\") pod \"nmstate-webhook-5f558f5558-z6585\" (UID: \"9fc4b697-3810-4744-8c3f-2285ee8399e8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 09:28:13 crc kubenswrapper[4930]: I0313 09:28:13.997533 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nsvm\" (UniqueName: \"kubernetes.io/projected/ca17bd84-571a-44a2-b882-1372cdc91ab2-kube-api-access-2nsvm\") pod \"nmstate-metrics-9b8c8685d-h8l9q\" (UID: \"ca17bd84-571a-44a2-b882-1372cdc91ab2\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-h8l9q" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.016779 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259"] Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.017758 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.021907 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.021955 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.022861 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-cvhzw" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.023867 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259"] Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.029322 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nsvm\" (UniqueName: \"kubernetes.io/projected/ca17bd84-571a-44a2-b882-1372cdc91ab2-kube-api-access-2nsvm\") pod \"nmstate-metrics-9b8c8685d-h8l9q\" (UID: \"ca17bd84-571a-44a2-b882-1372cdc91ab2\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-h8l9q" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.102299 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ab0b3cdd-cfe6-4d41-8406-18104b17c0fb-dbus-socket\") pod \"nmstate-handler-zrwjj\" (UID: \"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb\") " pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.102402 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9fc4b697-3810-4744-8c3f-2285ee8399e8-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-z6585\" (UID: \"9fc4b697-3810-4744-8c3f-2285ee8399e8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.102451 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bf8zh\" (UniqueName: \"kubernetes.io/projected/9fc4b697-3810-4744-8c3f-2285ee8399e8-kube-api-access-bf8zh\") pod \"nmstate-webhook-5f558f5558-z6585\" (UID: \"9fc4b697-3810-4744-8c3f-2285ee8399e8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.102474 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkx29\" (UniqueName: \"kubernetes.io/projected/ab0b3cdd-cfe6-4d41-8406-18104b17c0fb-kube-api-access-fkx29\") pod \"nmstate-handler-zrwjj\" (UID: \"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb\") " pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.102506 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ab0b3cdd-cfe6-4d41-8406-18104b17c0fb-nmstate-lock\") pod \"nmstate-handler-zrwjj\" (UID: \"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb\") " pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.102619 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ab0b3cdd-cfe6-4d41-8406-18104b17c0fb-ovs-socket\") pod \"nmstate-handler-zrwjj\" (UID: \"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb\") " pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: E0313 09:28:14.103872 4930 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Mar 13 09:28:14 crc kubenswrapper[4930]: E0313 09:28:14.103947 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9fc4b697-3810-4744-8c3f-2285ee8399e8-tls-key-pair podName:9fc4b697-3810-4744-8c3f-2285ee8399e8 nodeName:}" failed. No retries permitted until 2026-03-13 09:28:14.603925623 +0000 UTC m=+935.353840300 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/9fc4b697-3810-4744-8c3f-2285ee8399e8-tls-key-pair") pod "nmstate-webhook-5f558f5558-z6585" (UID: "9fc4b697-3810-4744-8c3f-2285ee8399e8") : secret "openshift-nmstate-webhook" not found Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.131186 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bf8zh\" (UniqueName: \"kubernetes.io/projected/9fc4b697-3810-4744-8c3f-2285ee8399e8-kube-api-access-bf8zh\") pod \"nmstate-webhook-5f558f5558-z6585\" (UID: \"9fc4b697-3810-4744-8c3f-2285ee8399e8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.200384 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-h8l9q" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.204588 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmg7j\" (UniqueName: \"kubernetes.io/projected/9bff675b-3745-485a-a492-d32038ded169-kube-api-access-pmg7j\") pod \"nmstate-console-plugin-86f58fcf4-58259\" (UID: \"9bff675b-3745-485a-a492-d32038ded169\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.204651 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ab0b3cdd-cfe6-4d41-8406-18104b17c0fb-dbus-socket\") pod \"nmstate-handler-zrwjj\" (UID: \"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb\") " pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.204817 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9bff675b-3745-485a-a492-d32038ded169-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-58259\" (UID: \"9bff675b-3745-485a-a492-d32038ded169\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.204875 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkx29\" (UniqueName: \"kubernetes.io/projected/ab0b3cdd-cfe6-4d41-8406-18104b17c0fb-kube-api-access-fkx29\") pod \"nmstate-handler-zrwjj\" (UID: \"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb\") " pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.204955 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/ab0b3cdd-cfe6-4d41-8406-18104b17c0fb-dbus-socket\") pod \"nmstate-handler-zrwjj\" (UID: \"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb\") " pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.204970 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ab0b3cdd-cfe6-4d41-8406-18104b17c0fb-nmstate-lock\") pod \"nmstate-handler-zrwjj\" (UID: \"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb\") " pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.205005 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/ab0b3cdd-cfe6-4d41-8406-18104b17c0fb-nmstate-lock\") pod \"nmstate-handler-zrwjj\" (UID: \"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb\") " pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.205072 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9bff675b-3745-485a-a492-d32038ded169-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-58259\" (UID: \"9bff675b-3745-485a-a492-d32038ded169\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.205092 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ab0b3cdd-cfe6-4d41-8406-18104b17c0fb-ovs-socket\") pod \"nmstate-handler-zrwjj\" (UID: \"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb\") " pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.205124 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/ab0b3cdd-cfe6-4d41-8406-18104b17c0fb-ovs-socket\") pod \"nmstate-handler-zrwjj\" (UID: \"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb\") " pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.209389 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7b97b57845-4zjvn"] Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.210594 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.232920 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7b97b57845-4zjvn"] Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.246219 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkx29\" (UniqueName: \"kubernetes.io/projected/ab0b3cdd-cfe6-4d41-8406-18104b17c0fb-kube-api-access-fkx29\") pod \"nmstate-handler-zrwjj\" (UID: \"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb\") " pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.306656 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-serving-cert\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.306928 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmg7j\" (UniqueName: \"kubernetes.io/projected/9bff675b-3745-485a-a492-d32038ded169-kube-api-access-pmg7j\") pod \"nmstate-console-plugin-86f58fcf4-58259\" (UID: \"9bff675b-3745-485a-a492-d32038ded169\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.306966 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-trusted-ca-bundle\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.306990 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-config\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.307038 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9bff675b-3745-485a-a492-d32038ded169-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-58259\" (UID: \"9bff675b-3745-485a-a492-d32038ded169\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.307095 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-oauth-serving-cert\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.307113 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5tb7\" (UniqueName: \"kubernetes.io/projected/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-kube-api-access-s5tb7\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.307144 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-service-ca\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.307163 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-oauth-config\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.307183 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9bff675b-3745-485a-a492-d32038ded169-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-58259\" (UID: \"9bff675b-3745-485a-a492-d32038ded169\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.308009 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/9bff675b-3745-485a-a492-d32038ded169-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-58259\" (UID: \"9bff675b-3745-485a-a492-d32038ded169\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.312249 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/9bff675b-3745-485a-a492-d32038ded169-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-58259\" (UID: \"9bff675b-3745-485a-a492-d32038ded169\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.324766 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmg7j\" (UniqueName: \"kubernetes.io/projected/9bff675b-3745-485a-a492-d32038ded169-kube-api-access-pmg7j\") pod \"nmstate-console-plugin-86f58fcf4-58259\" (UID: \"9bff675b-3745-485a-a492-d32038ded169\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.367942 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.409500 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-oauth-serving-cert\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.409546 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5tb7\" (UniqueName: \"kubernetes.io/projected/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-kube-api-access-s5tb7\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.409592 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-service-ca\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.409622 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-oauth-config\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.409682 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-serving-cert\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.409714 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-trusted-ca-bundle\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.409743 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-config\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.410548 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-oauth-serving-cert\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.411132 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-config\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.411515 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-service-ca\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.412312 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-trusted-ca-bundle\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.415575 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-oauth-config\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.416125 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-serving-cert\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.452952 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5tb7\" (UniqueName: \"kubernetes.io/projected/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-kube-api-access-s5tb7\") pod \"console-7b97b57845-4zjvn\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.531824 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.534725 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.622466 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9fc4b697-3810-4744-8c3f-2285ee8399e8-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-z6585\" (UID: \"9fc4b697-3810-4744-8c3f-2285ee8399e8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.627919 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/9fc4b697-3810-4744-8c3f-2285ee8399e8-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-z6585\" (UID: \"9fc4b697-3810-4744-8c3f-2285ee8399e8\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.726592 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-h8l9q"] Mar 13 09:28:14 crc kubenswrapper[4930]: W0313 09:28:14.753673 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca17bd84_571a_44a2_b882_1372cdc91ab2.slice/crio-123ff88d2fa4c35ab338b4096adfc1e586c942cf77b2005f83b1918c84dd49d5 WatchSource:0}: Error finding container 123ff88d2fa4c35ab338b4096adfc1e586c942cf77b2005f83b1918c84dd49d5: Status 404 returned error can't find the container with id 123ff88d2fa4c35ab338b4096adfc1e586c942cf77b2005f83b1918c84dd49d5 Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.779720 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259"] Mar 13 09:28:14 crc kubenswrapper[4930]: I0313 09:28:14.785319 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 09:28:15 crc kubenswrapper[4930]: I0313 09:28:15.070967 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-zrwjj" event={"ID":"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb","Type":"ContainerStarted","Data":"af48d78f3e3c535c4ca42537fc2d9476dc1f1415a8080b728c280fab0df76cd4"} Mar 13 09:28:15 crc kubenswrapper[4930]: I0313 09:28:15.072364 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-h8l9q" event={"ID":"ca17bd84-571a-44a2-b882-1372cdc91ab2","Type":"ContainerStarted","Data":"123ff88d2fa4c35ab338b4096adfc1e586c942cf77b2005f83b1918c84dd49d5"} Mar 13 09:28:15 crc kubenswrapper[4930]: I0313 09:28:15.072388 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7b97b57845-4zjvn"] Mar 13 09:28:15 crc kubenswrapper[4930]: W0313 09:28:15.073480 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7c5b562e_e1d3_4a2d_8169_e4b8a687251a.slice/crio-3fc82d44e00d139493a20dc6e1f6848aa1eaa78e449a44127ea3e5759d6edddc WatchSource:0}: Error finding container 3fc82d44e00d139493a20dc6e1f6848aa1eaa78e449a44127ea3e5759d6edddc: Status 404 returned error can't find the container with id 3fc82d44e00d139493a20dc6e1f6848aa1eaa78e449a44127ea3e5759d6edddc Mar 13 09:28:15 crc kubenswrapper[4930]: I0313 09:28:15.073527 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" event={"ID":"9bff675b-3745-485a-a492-d32038ded169","Type":"ContainerStarted","Data":"80dfc516148a04ce350c98cc848fc669dcd6c85563a3801c8c8f18492dcdab49"} Mar 13 09:28:15 crc kubenswrapper[4930]: I0313 09:28:15.214551 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-z6585"] Mar 13 09:28:16 crc kubenswrapper[4930]: I0313 09:28:16.081896 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7b97b57845-4zjvn" event={"ID":"7c5b562e-e1d3-4a2d-8169-e4b8a687251a","Type":"ContainerStarted","Data":"36e5fa63122da9c19f41e3bf274a7c37b2bb9f6a39f72e7e6bff6f02d133cb71"} Mar 13 09:28:16 crc kubenswrapper[4930]: I0313 09:28:16.081934 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7b97b57845-4zjvn" event={"ID":"7c5b562e-e1d3-4a2d-8169-e4b8a687251a","Type":"ContainerStarted","Data":"3fc82d44e00d139493a20dc6e1f6848aa1eaa78e449a44127ea3e5759d6edddc"} Mar 13 09:28:16 crc kubenswrapper[4930]: I0313 09:28:16.083651 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" event={"ID":"9fc4b697-3810-4744-8c3f-2285ee8399e8","Type":"ContainerStarted","Data":"4ce3fcf429d0a378d0f1c1a7c64b2e78c1e8461e23fd19b07e66e610ecb772a6"} Mar 13 09:28:16 crc kubenswrapper[4930]: I0313 09:28:16.103578 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7b97b57845-4zjvn" podStartSLOduration=2.103562056 podStartE2EDuration="2.103562056s" podCreationTimestamp="2026-03-13 09:28:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:28:16.102407227 +0000 UTC m=+936.852321904" watchObservedRunningTime="2026-03-13 09:28:16.103562056 +0000 UTC m=+936.853476733" Mar 13 09:28:18 crc kubenswrapper[4930]: I0313 09:28:18.108258 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-h8l9q" event={"ID":"ca17bd84-571a-44a2-b882-1372cdc91ab2","Type":"ContainerStarted","Data":"5f3644c0bb15fc1c6343c95dec7160516ce3df508bed0dfba583bd8aa2315907"} Mar 13 09:28:18 crc kubenswrapper[4930]: I0313 09:28:18.109896 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" event={"ID":"9fc4b697-3810-4744-8c3f-2285ee8399e8","Type":"ContainerStarted","Data":"5fadc37631430233ca7e11f2756939a1e6c29c16c135999e20784fb50ea52681"} Mar 13 09:28:18 crc kubenswrapper[4930]: I0313 09:28:18.110032 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 09:28:18 crc kubenswrapper[4930]: I0313 09:28:18.111376 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" event={"ID":"9bff675b-3745-485a-a492-d32038ded169","Type":"ContainerStarted","Data":"c37c8b561a81357a93533a3485b1e68eee4ee50edd7a72785ce0046d540f63dc"} Mar 13 09:28:18 crc kubenswrapper[4930]: I0313 09:28:18.114152 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-zrwjj" event={"ID":"ab0b3cdd-cfe6-4d41-8406-18104b17c0fb","Type":"ContainerStarted","Data":"98e0416b3195bb0891cf1ed5e413bb624993fd5878552530c93eb1605d3e7d00"} Mar 13 09:28:18 crc kubenswrapper[4930]: I0313 09:28:18.114557 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:18 crc kubenswrapper[4930]: I0313 09:28:18.130672 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" podStartSLOduration=2.636833804 podStartE2EDuration="5.13065244s" podCreationTimestamp="2026-03-13 09:28:13 +0000 UTC" firstStartedPulling="2026-03-13 09:28:15.227013626 +0000 UTC m=+935.976928303" lastFinishedPulling="2026-03-13 09:28:17.720832272 +0000 UTC m=+938.470746939" observedRunningTime="2026-03-13 09:28:18.123347909 +0000 UTC m=+938.873262586" watchObservedRunningTime="2026-03-13 09:28:18.13065244 +0000 UTC m=+938.880567117" Mar 13 09:28:18 crc kubenswrapper[4930]: I0313 09:28:18.151065 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-58259" podStartSLOduration=2.23883818 podStartE2EDuration="5.151040317s" podCreationTimestamp="2026-03-13 09:28:13 +0000 UTC" firstStartedPulling="2026-03-13 09:28:14.808592534 +0000 UTC m=+935.558507211" lastFinishedPulling="2026-03-13 09:28:17.720794671 +0000 UTC m=+938.470709348" observedRunningTime="2026-03-13 09:28:18.145395627 +0000 UTC m=+938.895310314" watchObservedRunningTime="2026-03-13 09:28:18.151040317 +0000 UTC m=+938.900954994" Mar 13 09:28:18 crc kubenswrapper[4930]: I0313 09:28:18.185928 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-zrwjj" podStartSLOduration=2.064839768 podStartE2EDuration="5.185908233s" podCreationTimestamp="2026-03-13 09:28:13 +0000 UTC" firstStartedPulling="2026-03-13 09:28:14.594722442 +0000 UTC m=+935.344637119" lastFinishedPulling="2026-03-13 09:28:17.715790917 +0000 UTC m=+938.465705584" observedRunningTime="2026-03-13 09:28:18.179975895 +0000 UTC m=+938.929890572" watchObservedRunningTime="2026-03-13 09:28:18.185908233 +0000 UTC m=+938.935822920" Mar 13 09:28:21 crc kubenswrapper[4930]: I0313 09:28:21.141191 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-h8l9q" event={"ID":"ca17bd84-571a-44a2-b882-1372cdc91ab2","Type":"ContainerStarted","Data":"cf2556bcf3681ec006fc2652add631a4a26166ac06fd9a9da32f3144e1b62e09"} Mar 13 09:28:21 crc kubenswrapper[4930]: I0313 09:28:21.165044 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-h8l9q" podStartSLOduration=2.140250021 podStartE2EDuration="8.165026062s" podCreationTimestamp="2026-03-13 09:28:13 +0000 UTC" firstStartedPulling="2026-03-13 09:28:14.76253395 +0000 UTC m=+935.512448627" lastFinishedPulling="2026-03-13 09:28:20.787309991 +0000 UTC m=+941.537224668" observedRunningTime="2026-03-13 09:28:21.159515665 +0000 UTC m=+941.909430362" watchObservedRunningTime="2026-03-13 09:28:21.165026062 +0000 UTC m=+941.914940729" Mar 13 09:28:24 crc kubenswrapper[4930]: I0313 09:28:24.532786 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:24 crc kubenswrapper[4930]: I0313 09:28:24.533146 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:24 crc kubenswrapper[4930]: I0313 09:28:24.542014 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:24 crc kubenswrapper[4930]: I0313 09:28:24.570628 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 09:28:25 crc kubenswrapper[4930]: I0313 09:28:25.184901 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:28:25 crc kubenswrapper[4930]: I0313 09:28:25.244859 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-767c6dcbb4-wplpt"] Mar 13 09:28:34 crc kubenswrapper[4930]: I0313 09:28:34.791199 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 09:28:41 crc kubenswrapper[4930]: I0313 09:28:41.496188 4930 scope.go:117] "RemoveContainer" containerID="d9a2fddfc4f8d853db63aaa1a5552cac9402357e9ef7eadccc91d7a533b081e3" Mar 13 09:28:42 crc kubenswrapper[4930]: I0313 09:28:42.308896 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:28:42 crc kubenswrapper[4930]: I0313 09:28:42.309309 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.299068 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-767c6dcbb4-wplpt" podUID="4255a8cb-2916-4af4-b84d-7c5650efb80b" containerName="console" containerID="cri-o://7604123992f68375b2ae37007af141f5a622bd2bb74f40a7ebf748f175739864" gracePeriod=15 Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.629032 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-767c6dcbb4-wplpt_4255a8cb-2916-4af4-b84d-7c5650efb80b/console/0.log" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.629307 4930 generic.go:334] "Generic (PLEG): container finished" podID="4255a8cb-2916-4af4-b84d-7c5650efb80b" containerID="7604123992f68375b2ae37007af141f5a622bd2bb74f40a7ebf748f175739864" exitCode=2 Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.629337 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-767c6dcbb4-wplpt" event={"ID":"4255a8cb-2916-4af4-b84d-7c5650efb80b","Type":"ContainerDied","Data":"7604123992f68375b2ae37007af141f5a622bd2bb74f40a7ebf748f175739864"} Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.629359 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-767c6dcbb4-wplpt" event={"ID":"4255a8cb-2916-4af4-b84d-7c5650efb80b","Type":"ContainerDied","Data":"ba7d89e82a6f2aedd026af5efab6112cce0af5cff44c3aec84bc72c7fb32d69c"} Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.629370 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ba7d89e82a6f2aedd026af5efab6112cce0af5cff44c3aec84bc72c7fb32d69c" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.663953 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-767c6dcbb4-wplpt_4255a8cb-2916-4af4-b84d-7c5650efb80b/console/0.log" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.664015 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.856190 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-trusted-ca-bundle\") pod \"4255a8cb-2916-4af4-b84d-7c5650efb80b\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.856239 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-oauth-config\") pod \"4255a8cb-2916-4af4-b84d-7c5650efb80b\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.856307 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-serving-cert\") pod \"4255a8cb-2916-4af4-b84d-7c5650efb80b\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.856346 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-oauth-serving-cert\") pod \"4255a8cb-2916-4af4-b84d-7c5650efb80b\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.856409 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-config\") pod \"4255a8cb-2916-4af4-b84d-7c5650efb80b\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.856426 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-service-ca\") pod \"4255a8cb-2916-4af4-b84d-7c5650efb80b\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.856456 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ddwc\" (UniqueName: \"kubernetes.io/projected/4255a8cb-2916-4af4-b84d-7c5650efb80b-kube-api-access-7ddwc\") pod \"4255a8cb-2916-4af4-b84d-7c5650efb80b\" (UID: \"4255a8cb-2916-4af4-b84d-7c5650efb80b\") " Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.857870 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "4255a8cb-2916-4af4-b84d-7c5650efb80b" (UID: "4255a8cb-2916-4af4-b84d-7c5650efb80b"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.857896 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "4255a8cb-2916-4af4-b84d-7c5650efb80b" (UID: "4255a8cb-2916-4af4-b84d-7c5650efb80b"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.857951 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-config" (OuterVolumeSpecName: "console-config") pod "4255a8cb-2916-4af4-b84d-7c5650efb80b" (UID: "4255a8cb-2916-4af4-b84d-7c5650efb80b"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.857963 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-service-ca" (OuterVolumeSpecName: "service-ca") pod "4255a8cb-2916-4af4-b84d-7c5650efb80b" (UID: "4255a8cb-2916-4af4-b84d-7c5650efb80b"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.862383 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "4255a8cb-2916-4af4-b84d-7c5650efb80b" (UID: "4255a8cb-2916-4af4-b84d-7c5650efb80b"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.863868 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4255a8cb-2916-4af4-b84d-7c5650efb80b-kube-api-access-7ddwc" (OuterVolumeSpecName: "kube-api-access-7ddwc") pod "4255a8cb-2916-4af4-b84d-7c5650efb80b" (UID: "4255a8cb-2916-4af4-b84d-7c5650efb80b"). InnerVolumeSpecName "kube-api-access-7ddwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.864700 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "4255a8cb-2916-4af4-b84d-7c5650efb80b" (UID: "4255a8cb-2916-4af4-b84d-7c5650efb80b"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.958134 4930 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.958181 4930 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.958191 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ddwc\" (UniqueName: \"kubernetes.io/projected/4255a8cb-2916-4af4-b84d-7c5650efb80b-kube-api-access-7ddwc\") on node \"crc\" DevicePath \"\"" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.958208 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.958220 4930 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.958229 4930 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/4255a8cb-2916-4af4-b84d-7c5650efb80b-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:28:50 crc kubenswrapper[4930]: I0313 09:28:50.958242 4930 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/4255a8cb-2916-4af4-b84d-7c5650efb80b-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:28:51 crc kubenswrapper[4930]: I0313 09:28:51.636553 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-767c6dcbb4-wplpt" Mar 13 09:28:51 crc kubenswrapper[4930]: I0313 09:28:51.668642 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-767c6dcbb4-wplpt"] Mar 13 09:28:51 crc kubenswrapper[4930]: I0313 09:28:51.675322 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-767c6dcbb4-wplpt"] Mar 13 09:28:51 crc kubenswrapper[4930]: I0313 09:28:51.978502 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4255a8cb-2916-4af4-b84d-7c5650efb80b" path="/var/lib/kubelet/pods/4255a8cb-2916-4af4-b84d-7c5650efb80b/volumes" Mar 13 09:28:53 crc kubenswrapper[4930]: I0313 09:28:53.839644 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8"] Mar 13 09:28:53 crc kubenswrapper[4930]: E0313 09:28:53.840174 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4255a8cb-2916-4af4-b84d-7c5650efb80b" containerName="console" Mar 13 09:28:53 crc kubenswrapper[4930]: I0313 09:28:53.840186 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="4255a8cb-2916-4af4-b84d-7c5650efb80b" containerName="console" Mar 13 09:28:53 crc kubenswrapper[4930]: I0313 09:28:53.840322 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="4255a8cb-2916-4af4-b84d-7c5650efb80b" containerName="console" Mar 13 09:28:53 crc kubenswrapper[4930]: I0313 09:28:53.841250 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:28:53 crc kubenswrapper[4930]: I0313 09:28:53.855610 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 13 09:28:53 crc kubenswrapper[4930]: I0313 09:28:53.856634 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8"] Mar 13 09:28:53 crc kubenswrapper[4930]: I0313 09:28:53.898422 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8rqx\" (UniqueName: \"kubernetes.io/projected/41196fbb-c978-4131-93e1-b572cde9aebf-kube-api-access-k8rqx\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8\" (UID: \"41196fbb-c978-4131-93e1-b572cde9aebf\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:28:53 crc kubenswrapper[4930]: I0313 09:28:53.898548 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41196fbb-c978-4131-93e1-b572cde9aebf-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8\" (UID: \"41196fbb-c978-4131-93e1-b572cde9aebf\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:28:53 crc kubenswrapper[4930]: I0313 09:28:53.898688 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41196fbb-c978-4131-93e1-b572cde9aebf-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8\" (UID: \"41196fbb-c978-4131-93e1-b572cde9aebf\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:28:53 crc kubenswrapper[4930]: I0313 09:28:53.999740 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41196fbb-c978-4131-93e1-b572cde9aebf-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8\" (UID: \"41196fbb-c978-4131-93e1-b572cde9aebf\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:28:53 crc kubenswrapper[4930]: I0313 09:28:53.999812 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41196fbb-c978-4131-93e1-b572cde9aebf-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8\" (UID: \"41196fbb-c978-4131-93e1-b572cde9aebf\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:28:53 crc kubenswrapper[4930]: I0313 09:28:53.999889 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8rqx\" (UniqueName: \"kubernetes.io/projected/41196fbb-c978-4131-93e1-b572cde9aebf-kube-api-access-k8rqx\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8\" (UID: \"41196fbb-c978-4131-93e1-b572cde9aebf\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:28:54 crc kubenswrapper[4930]: I0313 09:28:54.000427 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41196fbb-c978-4131-93e1-b572cde9aebf-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8\" (UID: \"41196fbb-c978-4131-93e1-b572cde9aebf\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:28:54 crc kubenswrapper[4930]: I0313 09:28:54.000581 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41196fbb-c978-4131-93e1-b572cde9aebf-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8\" (UID: \"41196fbb-c978-4131-93e1-b572cde9aebf\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:28:54 crc kubenswrapper[4930]: I0313 09:28:54.015759 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8rqx\" (UniqueName: \"kubernetes.io/projected/41196fbb-c978-4131-93e1-b572cde9aebf-kube-api-access-k8rqx\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8\" (UID: \"41196fbb-c978-4131-93e1-b572cde9aebf\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:28:54 crc kubenswrapper[4930]: I0313 09:28:54.159429 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:28:54 crc kubenswrapper[4930]: I0313 09:28:54.565601 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8"] Mar 13 09:28:54 crc kubenswrapper[4930]: I0313 09:28:54.657642 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" event={"ID":"41196fbb-c978-4131-93e1-b572cde9aebf","Type":"ContainerStarted","Data":"2b75567d5e05b0ae2f5d549ada122b60850446f160a61c2ecdde035e454b3bbd"} Mar 13 09:28:55 crc kubenswrapper[4930]: I0313 09:28:55.666324 4930 generic.go:334] "Generic (PLEG): container finished" podID="41196fbb-c978-4131-93e1-b572cde9aebf" containerID="15d9c2e0c7c6b85a9f93312f03c105c9bd6b0bb303eb4dd05397c19195fab659" exitCode=0 Mar 13 09:28:55 crc kubenswrapper[4930]: I0313 09:28:55.666380 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" event={"ID":"41196fbb-c978-4131-93e1-b572cde9aebf","Type":"ContainerDied","Data":"15d9c2e0c7c6b85a9f93312f03c105c9bd6b0bb303eb4dd05397c19195fab659"} Mar 13 09:28:58 crc kubenswrapper[4930]: I0313 09:28:58.689376 4930 generic.go:334] "Generic (PLEG): container finished" podID="41196fbb-c978-4131-93e1-b572cde9aebf" containerID="f70e4fa1d2f7fe6604be24bbf47dac57654c8837dd22a350a822df3fa642b06f" exitCode=0 Mar 13 09:28:58 crc kubenswrapper[4930]: I0313 09:28:58.689408 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" event={"ID":"41196fbb-c978-4131-93e1-b572cde9aebf","Type":"ContainerDied","Data":"f70e4fa1d2f7fe6604be24bbf47dac57654c8837dd22a350a822df3fa642b06f"} Mar 13 09:28:59 crc kubenswrapper[4930]: I0313 09:28:59.700277 4930 generic.go:334] "Generic (PLEG): container finished" podID="41196fbb-c978-4131-93e1-b572cde9aebf" containerID="2546b5966fd9f44deee7b9ea127b5167cac8429bfe55723d9471d76646019714" exitCode=0 Mar 13 09:28:59 crc kubenswrapper[4930]: I0313 09:28:59.700317 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" event={"ID":"41196fbb-c978-4131-93e1-b572cde9aebf","Type":"ContainerDied","Data":"2546b5966fd9f44deee7b9ea127b5167cac8429bfe55723d9471d76646019714"} Mar 13 09:29:00 crc kubenswrapper[4930]: I0313 09:29:00.963688 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:29:01 crc kubenswrapper[4930]: I0313 09:29:01.152709 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8rqx\" (UniqueName: \"kubernetes.io/projected/41196fbb-c978-4131-93e1-b572cde9aebf-kube-api-access-k8rqx\") pod \"41196fbb-c978-4131-93e1-b572cde9aebf\" (UID: \"41196fbb-c978-4131-93e1-b572cde9aebf\") " Mar 13 09:29:01 crc kubenswrapper[4930]: I0313 09:29:01.152824 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41196fbb-c978-4131-93e1-b572cde9aebf-bundle\") pod \"41196fbb-c978-4131-93e1-b572cde9aebf\" (UID: \"41196fbb-c978-4131-93e1-b572cde9aebf\") " Mar 13 09:29:01 crc kubenswrapper[4930]: I0313 09:29:01.152877 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41196fbb-c978-4131-93e1-b572cde9aebf-util\") pod \"41196fbb-c978-4131-93e1-b572cde9aebf\" (UID: \"41196fbb-c978-4131-93e1-b572cde9aebf\") " Mar 13 09:29:01 crc kubenswrapper[4930]: I0313 09:29:01.154268 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41196fbb-c978-4131-93e1-b572cde9aebf-bundle" (OuterVolumeSpecName: "bundle") pod "41196fbb-c978-4131-93e1-b572cde9aebf" (UID: "41196fbb-c978-4131-93e1-b572cde9aebf"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:29:01 crc kubenswrapper[4930]: I0313 09:29:01.158519 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41196fbb-c978-4131-93e1-b572cde9aebf-kube-api-access-k8rqx" (OuterVolumeSpecName: "kube-api-access-k8rqx") pod "41196fbb-c978-4131-93e1-b572cde9aebf" (UID: "41196fbb-c978-4131-93e1-b572cde9aebf"). InnerVolumeSpecName "kube-api-access-k8rqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:29:01 crc kubenswrapper[4930]: I0313 09:29:01.163261 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41196fbb-c978-4131-93e1-b572cde9aebf-util" (OuterVolumeSpecName: "util") pod "41196fbb-c978-4131-93e1-b572cde9aebf" (UID: "41196fbb-c978-4131-93e1-b572cde9aebf"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:29:01 crc kubenswrapper[4930]: I0313 09:29:01.255218 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8rqx\" (UniqueName: \"kubernetes.io/projected/41196fbb-c978-4131-93e1-b572cde9aebf-kube-api-access-k8rqx\") on node \"crc\" DevicePath \"\"" Mar 13 09:29:01 crc kubenswrapper[4930]: I0313 09:29:01.255280 4930 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/41196fbb-c978-4131-93e1-b572cde9aebf-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:29:01 crc kubenswrapper[4930]: I0313 09:29:01.255298 4930 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/41196fbb-c978-4131-93e1-b572cde9aebf-util\") on node \"crc\" DevicePath \"\"" Mar 13 09:29:01 crc kubenswrapper[4930]: I0313 09:29:01.721194 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" event={"ID":"41196fbb-c978-4131-93e1-b572cde9aebf","Type":"ContainerDied","Data":"2b75567d5e05b0ae2f5d549ada122b60850446f160a61c2ecdde035e454b3bbd"} Mar 13 09:29:01 crc kubenswrapper[4930]: I0313 09:29:01.721261 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c17hdk8" Mar 13 09:29:01 crc kubenswrapper[4930]: I0313 09:29:01.721280 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b75567d5e05b0ae2f5d549ada122b60850446f160a61c2ecdde035e454b3bbd" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.736714 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5"] Mar 13 09:29:11 crc kubenswrapper[4930]: E0313 09:29:11.737523 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41196fbb-c978-4131-93e1-b572cde9aebf" containerName="extract" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.737537 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="41196fbb-c978-4131-93e1-b572cde9aebf" containerName="extract" Mar 13 09:29:11 crc kubenswrapper[4930]: E0313 09:29:11.737548 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41196fbb-c978-4131-93e1-b572cde9aebf" containerName="pull" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.737554 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="41196fbb-c978-4131-93e1-b572cde9aebf" containerName="pull" Mar 13 09:29:11 crc kubenswrapper[4930]: E0313 09:29:11.737561 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41196fbb-c978-4131-93e1-b572cde9aebf" containerName="util" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.737568 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="41196fbb-c978-4131-93e1-b572cde9aebf" containerName="util" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.737698 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="41196fbb-c978-4131-93e1-b572cde9aebf" containerName="extract" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.738228 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.740136 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.740939 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.741819 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.741922 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.742134 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-pdzfz" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.755304 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5"] Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.940316 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a-apiservice-cert\") pod \"metallb-operator-controller-manager-5d4f94cfb6-bd6n5\" (UID: \"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a\") " pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.940378 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8hv8\" (UniqueName: \"kubernetes.io/projected/cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a-kube-api-access-r8hv8\") pod \"metallb-operator-controller-manager-5d4f94cfb6-bd6n5\" (UID: \"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a\") " pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:11 crc kubenswrapper[4930]: I0313 09:29:11.940495 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a-webhook-cert\") pod \"metallb-operator-controller-manager-5d4f94cfb6-bd6n5\" (UID: \"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a\") " pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.041586 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a-apiservice-cert\") pod \"metallb-operator-controller-manager-5d4f94cfb6-bd6n5\" (UID: \"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a\") " pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.041634 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8hv8\" (UniqueName: \"kubernetes.io/projected/cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a-kube-api-access-r8hv8\") pod \"metallb-operator-controller-manager-5d4f94cfb6-bd6n5\" (UID: \"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a\") " pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.041665 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a-webhook-cert\") pod \"metallb-operator-controller-manager-5d4f94cfb6-bd6n5\" (UID: \"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a\") " pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.049047 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a-webhook-cert\") pod \"metallb-operator-controller-manager-5d4f94cfb6-bd6n5\" (UID: \"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a\") " pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.049455 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a-apiservice-cert\") pod \"metallb-operator-controller-manager-5d4f94cfb6-bd6n5\" (UID: \"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a\") " pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.089914 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8hv8\" (UniqueName: \"kubernetes.io/projected/cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a-kube-api-access-r8hv8\") pod \"metallb-operator-controller-manager-5d4f94cfb6-bd6n5\" (UID: \"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a\") " pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.133080 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5"] Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.134214 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.136314 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-f8sfl" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.136754 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.138887 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.143581 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f338c47b-e688-4de8-8bd7-02be68a70ad3-apiservice-cert\") pod \"metallb-operator-webhook-server-df66cd457-h9hb5\" (UID: \"f338c47b-e688-4de8-8bd7-02be68a70ad3\") " pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.143731 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f338c47b-e688-4de8-8bd7-02be68a70ad3-webhook-cert\") pod \"metallb-operator-webhook-server-df66cd457-h9hb5\" (UID: \"f338c47b-e688-4de8-8bd7-02be68a70ad3\") " pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.143790 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxfbz\" (UniqueName: \"kubernetes.io/projected/f338c47b-e688-4de8-8bd7-02be68a70ad3-kube-api-access-pxfbz\") pod \"metallb-operator-webhook-server-df66cd457-h9hb5\" (UID: \"f338c47b-e688-4de8-8bd7-02be68a70ad3\") " pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.171623 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5"] Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.245149 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f338c47b-e688-4de8-8bd7-02be68a70ad3-apiservice-cert\") pod \"metallb-operator-webhook-server-df66cd457-h9hb5\" (UID: \"f338c47b-e688-4de8-8bd7-02be68a70ad3\") " pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.245235 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f338c47b-e688-4de8-8bd7-02be68a70ad3-webhook-cert\") pod \"metallb-operator-webhook-server-df66cd457-h9hb5\" (UID: \"f338c47b-e688-4de8-8bd7-02be68a70ad3\") " pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.245288 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxfbz\" (UniqueName: \"kubernetes.io/projected/f338c47b-e688-4de8-8bd7-02be68a70ad3-kube-api-access-pxfbz\") pod \"metallb-operator-webhook-server-df66cd457-h9hb5\" (UID: \"f338c47b-e688-4de8-8bd7-02be68a70ad3\") " pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.249379 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f338c47b-e688-4de8-8bd7-02be68a70ad3-apiservice-cert\") pod \"metallb-operator-webhook-server-df66cd457-h9hb5\" (UID: \"f338c47b-e688-4de8-8bd7-02be68a70ad3\") " pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.262008 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f338c47b-e688-4de8-8bd7-02be68a70ad3-webhook-cert\") pod \"metallb-operator-webhook-server-df66cd457-h9hb5\" (UID: \"f338c47b-e688-4de8-8bd7-02be68a70ad3\") " pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.272194 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxfbz\" (UniqueName: \"kubernetes.io/projected/f338c47b-e688-4de8-8bd7-02be68a70ad3-kube-api-access-pxfbz\") pod \"metallb-operator-webhook-server-df66cd457-h9hb5\" (UID: \"f338c47b-e688-4de8-8bd7-02be68a70ad3\") " pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.308151 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.308207 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.308265 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.308989 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2a734be71e61d016548cb1b09f6ff35d98f17d0ce5f81f146f9d61495866a932"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.309053 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://2a734be71e61d016548cb1b09f6ff35d98f17d0ce5f81f146f9d61495866a932" gracePeriod=600 Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.353990 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.479928 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.810385 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="2a734be71e61d016548cb1b09f6ff35d98f17d0ce5f81f146f9d61495866a932" exitCode=0 Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.810524 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"2a734be71e61d016548cb1b09f6ff35d98f17d0ce5f81f146f9d61495866a932"} Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.810753 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"c06f278e05cf4208d0a63eab8c9e20f18523a288ed0cb0082dfaac4bd5791ab7"} Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.810798 4930 scope.go:117] "RemoveContainer" containerID="e32155b41a79cf2a274b0699cb480a931ddd70ce47fbf5a1faf4047a087b2d60" Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.826531 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5"] Mar 13 09:29:12 crc kubenswrapper[4930]: I0313 09:29:12.979774 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5"] Mar 13 09:29:12 crc kubenswrapper[4930]: W0313 09:29:12.988834 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf338c47b_e688_4de8_8bd7_02be68a70ad3.slice/crio-6b619ddee5b2f32e410a2e48722c2c8744fee1239e538f0ee02bc2d7de2cb59e WatchSource:0}: Error finding container 6b619ddee5b2f32e410a2e48722c2c8744fee1239e538f0ee02bc2d7de2cb59e: Status 404 returned error can't find the container with id 6b619ddee5b2f32e410a2e48722c2c8744fee1239e538f0ee02bc2d7de2cb59e Mar 13 09:29:13 crc kubenswrapper[4930]: I0313 09:29:13.820017 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" event={"ID":"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a","Type":"ContainerStarted","Data":"b3000c5f57a230eab549f6b9a0033d82003b387c0bd515a9d96c335ef3ecae32"} Mar 13 09:29:13 crc kubenswrapper[4930]: I0313 09:29:13.821098 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" event={"ID":"f338c47b-e688-4de8-8bd7-02be68a70ad3","Type":"ContainerStarted","Data":"6b619ddee5b2f32e410a2e48722c2c8744fee1239e538f0ee02bc2d7de2cb59e"} Mar 13 09:29:21 crc kubenswrapper[4930]: I0313 09:29:21.885017 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" event={"ID":"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a","Type":"ContainerStarted","Data":"1a6b7aecefef7fd19618182a43f445db9ec8c99bb1dc64ca39de9f5780dc2467"} Mar 13 09:29:21 crc kubenswrapper[4930]: I0313 09:29:21.885630 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:21 crc kubenswrapper[4930]: I0313 09:29:21.887340 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" event={"ID":"f338c47b-e688-4de8-8bd7-02be68a70ad3","Type":"ContainerStarted","Data":"9130180554e9bb6c5586aab89960a91bc9711671a3b4b502faee937ff892eda1"} Mar 13 09:29:21 crc kubenswrapper[4930]: I0313 09:29:21.887570 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:21 crc kubenswrapper[4930]: I0313 09:29:21.928711 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" podStartSLOduration=2.681994807 podStartE2EDuration="10.928687387s" podCreationTimestamp="2026-03-13 09:29:11 +0000 UTC" firstStartedPulling="2026-03-13 09:29:12.851115381 +0000 UTC m=+993.601030058" lastFinishedPulling="2026-03-13 09:29:21.097807951 +0000 UTC m=+1001.847722638" observedRunningTime="2026-03-13 09:29:21.92517606 +0000 UTC m=+1002.675090757" watchObservedRunningTime="2026-03-13 09:29:21.928687387 +0000 UTC m=+1002.678602064" Mar 13 09:29:21 crc kubenswrapper[4930]: I0313 09:29:21.953648 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" podStartSLOduration=1.8248280970000001 podStartE2EDuration="9.953629159s" podCreationTimestamp="2026-03-13 09:29:12 +0000 UTC" firstStartedPulling="2026-03-13 09:29:12.992170376 +0000 UTC m=+993.742085053" lastFinishedPulling="2026-03-13 09:29:21.120971428 +0000 UTC m=+1001.870886115" observedRunningTime="2026-03-13 09:29:21.94803625 +0000 UTC m=+1002.697950917" watchObservedRunningTime="2026-03-13 09:29:21.953629159 +0000 UTC m=+1002.703543836" Mar 13 09:29:32 crc kubenswrapper[4930]: I0313 09:29:32.487547 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 09:29:41 crc kubenswrapper[4930]: I0313 09:29:41.547360 4930 scope.go:117] "RemoveContainer" containerID="7604123992f68375b2ae37007af141f5a622bd2bb74f40a7ebf748f175739864" Mar 13 09:29:52 crc kubenswrapper[4930]: I0313 09:29:52.357186 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.125031 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx"] Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.126616 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.133911 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-nmrdx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.134089 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.147518 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-krmnx"] Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.151212 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.152463 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx"] Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.155879 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.156107 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.205625 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bk6gm\" (UniqueName: \"kubernetes.io/projected/bbf3e078-1f66-4bcc-a742-0835afcb3e79-kube-api-access-bk6gm\") pod \"frr-k8s-webhook-server-bcc4b6f68-2tktx\" (UID: \"bbf3e078-1f66-4bcc-a742-0835afcb3e79\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.205682 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1e87ea83-860c-469b-93ba-1f9cbc7db958-metrics\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.205710 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1e87ea83-860c-469b-93ba-1f9cbc7db958-frr-startup\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.205770 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1e87ea83-860c-469b-93ba-1f9cbc7db958-frr-conf\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.205792 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1e87ea83-860c-469b-93ba-1f9cbc7db958-reloader\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.205817 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bbf3e078-1f66-4bcc-a742-0835afcb3e79-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-2tktx\" (UID: \"bbf3e078-1f66-4bcc-a742-0835afcb3e79\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.205899 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1e87ea83-860c-469b-93ba-1f9cbc7db958-frr-sockets\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.205951 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e87ea83-860c-469b-93ba-1f9cbc7db958-metrics-certs\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.205986 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7md6\" (UniqueName: \"kubernetes.io/projected/1e87ea83-860c-469b-93ba-1f9cbc7db958-kube-api-access-s7md6\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.222675 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-hc6zf"] Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.229493 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-9bhdc"] Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.230864 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.232193 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hc6zf" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.238804 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.238907 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.238951 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.238986 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-fh854" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.239035 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.245371 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-9bhdc"] Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.307292 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1e87ea83-860c-469b-93ba-1f9cbc7db958-frr-conf\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.307581 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1e87ea83-860c-469b-93ba-1f9cbc7db958-reloader\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.307660 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bbf3e078-1f66-4bcc-a742-0835afcb3e79-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-2tktx\" (UID: \"bbf3e078-1f66-4bcc-a742-0835afcb3e79\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.307783 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1e87ea83-860c-469b-93ba-1f9cbc7db958-frr-sockets\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.307876 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e87ea83-860c-469b-93ba-1f9cbc7db958-metrics-certs\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.307961 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7md6\" (UniqueName: \"kubernetes.io/projected/1e87ea83-860c-469b-93ba-1f9cbc7db958-kube-api-access-s7md6\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.308051 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bk6gm\" (UniqueName: \"kubernetes.io/projected/bbf3e078-1f66-4bcc-a742-0835afcb3e79-kube-api-access-bk6gm\") pod \"frr-k8s-webhook-server-bcc4b6f68-2tktx\" (UID: \"bbf3e078-1f66-4bcc-a742-0835afcb3e79\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.308137 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1e87ea83-860c-469b-93ba-1f9cbc7db958-metrics\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.308232 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1e87ea83-860c-469b-93ba-1f9cbc7db958-frr-startup\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.309113 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/1e87ea83-860c-469b-93ba-1f9cbc7db958-frr-startup\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.309848 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/1e87ea83-860c-469b-93ba-1f9cbc7db958-frr-conf\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.310117 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/1e87ea83-860c-469b-93ba-1f9cbc7db958-reloader\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: E0313 09:29:53.310254 4930 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Mar 13 09:29:53 crc kubenswrapper[4930]: E0313 09:29:53.310374 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/bbf3e078-1f66-4bcc-a742-0835afcb3e79-cert podName:bbf3e078-1f66-4bcc-a742-0835afcb3e79 nodeName:}" failed. No retries permitted until 2026-03-13 09:29:53.810359398 +0000 UTC m=+1034.560274075 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/bbf3e078-1f66-4bcc-a742-0835afcb3e79-cert") pod "frr-k8s-webhook-server-bcc4b6f68-2tktx" (UID: "bbf3e078-1f66-4bcc-a742-0835afcb3e79") : secret "frr-k8s-webhook-server-cert" not found Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.310857 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/1e87ea83-860c-469b-93ba-1f9cbc7db958-frr-sockets\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.312099 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/1e87ea83-860c-469b-93ba-1f9cbc7db958-metrics\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.316364 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1e87ea83-860c-469b-93ba-1f9cbc7db958-metrics-certs\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.333319 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7md6\" (UniqueName: \"kubernetes.io/projected/1e87ea83-860c-469b-93ba-1f9cbc7db958-kube-api-access-s7md6\") pod \"frr-k8s-krmnx\" (UID: \"1e87ea83-860c-469b-93ba-1f9cbc7db958\") " pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.334217 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bk6gm\" (UniqueName: \"kubernetes.io/projected/bbf3e078-1f66-4bcc-a742-0835afcb3e79-kube-api-access-bk6gm\") pod \"frr-k8s-webhook-server-bcc4b6f68-2tktx\" (UID: \"bbf3e078-1f66-4bcc-a742-0835afcb3e79\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.410935 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8h2n\" (UniqueName: \"kubernetes.io/projected/aa56d2a5-5b3a-487c-9248-21ef5689562f-kube-api-access-r8h2n\") pod \"controller-7bb4cc7c98-9bhdc\" (UID: \"aa56d2a5-5b3a-487c-9248-21ef5689562f\") " pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.412089 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f1370aa-666b-439e-9411-ec5e70360091-metrics-certs\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.412241 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7f1370aa-666b-439e-9411-ec5e70360091-memberlist\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.412382 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt7dq\" (UniqueName: \"kubernetes.io/projected/7f1370aa-666b-439e-9411-ec5e70360091-kube-api-access-nt7dq\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.412506 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7f1370aa-666b-439e-9411-ec5e70360091-metallb-excludel2\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.412619 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa56d2a5-5b3a-487c-9248-21ef5689562f-metrics-certs\") pod \"controller-7bb4cc7c98-9bhdc\" (UID: \"aa56d2a5-5b3a-487c-9248-21ef5689562f\") " pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.412714 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa56d2a5-5b3a-487c-9248-21ef5689562f-cert\") pod \"controller-7bb4cc7c98-9bhdc\" (UID: \"aa56d2a5-5b3a-487c-9248-21ef5689562f\") " pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.479578 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-krmnx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.514477 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f1370aa-666b-439e-9411-ec5e70360091-metrics-certs\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.514546 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7f1370aa-666b-439e-9411-ec5e70360091-memberlist\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.514616 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt7dq\" (UniqueName: \"kubernetes.io/projected/7f1370aa-666b-439e-9411-ec5e70360091-kube-api-access-nt7dq\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.514636 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7f1370aa-666b-439e-9411-ec5e70360091-metallb-excludel2\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.514654 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa56d2a5-5b3a-487c-9248-21ef5689562f-metrics-certs\") pod \"controller-7bb4cc7c98-9bhdc\" (UID: \"aa56d2a5-5b3a-487c-9248-21ef5689562f\") " pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.514673 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa56d2a5-5b3a-487c-9248-21ef5689562f-cert\") pod \"controller-7bb4cc7c98-9bhdc\" (UID: \"aa56d2a5-5b3a-487c-9248-21ef5689562f\") " pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.514710 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8h2n\" (UniqueName: \"kubernetes.io/projected/aa56d2a5-5b3a-487c-9248-21ef5689562f-kube-api-access-r8h2n\") pod \"controller-7bb4cc7c98-9bhdc\" (UID: \"aa56d2a5-5b3a-487c-9248-21ef5689562f\") " pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:29:53 crc kubenswrapper[4930]: E0313 09:29:53.515237 4930 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 13 09:29:53 crc kubenswrapper[4930]: E0313 09:29:53.515324 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f1370aa-666b-439e-9411-ec5e70360091-memberlist podName:7f1370aa-666b-439e-9411-ec5e70360091 nodeName:}" failed. No retries permitted until 2026-03-13 09:29:54.015304056 +0000 UTC m=+1034.765218803 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7f1370aa-666b-439e-9411-ec5e70360091-memberlist") pod "speaker-hc6zf" (UID: "7f1370aa-666b-439e-9411-ec5e70360091") : secret "metallb-memberlist" not found Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.516154 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/7f1370aa-666b-439e-9411-ec5e70360091-metallb-excludel2\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.519739 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7f1370aa-666b-439e-9411-ec5e70360091-metrics-certs\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.521635 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/aa56d2a5-5b3a-487c-9248-21ef5689562f-metrics-certs\") pod \"controller-7bb4cc7c98-9bhdc\" (UID: \"aa56d2a5-5b3a-487c-9248-21ef5689562f\") " pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.528662 4930 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.534784 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8h2n\" (UniqueName: \"kubernetes.io/projected/aa56d2a5-5b3a-487c-9248-21ef5689562f-kube-api-access-r8h2n\") pod \"controller-7bb4cc7c98-9bhdc\" (UID: \"aa56d2a5-5b3a-487c-9248-21ef5689562f\") " pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.534969 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt7dq\" (UniqueName: \"kubernetes.io/projected/7f1370aa-666b-439e-9411-ec5e70360091-kube-api-access-nt7dq\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.535164 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/aa56d2a5-5b3a-487c-9248-21ef5689562f-cert\") pod \"controller-7bb4cc7c98-9bhdc\" (UID: \"aa56d2a5-5b3a-487c-9248-21ef5689562f\") " pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.554693 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.821122 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bbf3e078-1f66-4bcc-a742-0835afcb3e79-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-2tktx\" (UID: \"bbf3e078-1f66-4bcc-a742-0835afcb3e79\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.825516 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/bbf3e078-1f66-4bcc-a742-0835afcb3e79-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-2tktx\" (UID: \"bbf3e078-1f66-4bcc-a742-0835afcb3e79\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 09:29:53 crc kubenswrapper[4930]: I0313 09:29:53.963661 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-9bhdc"] Mar 13 09:29:53 crc kubenswrapper[4930]: W0313 09:29:53.968647 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaa56d2a5_5b3a_487c_9248_21ef5689562f.slice/crio-2ae143e07b24e9b0cb1ed0b5cb463bef5b8b0f1aeae1b1118eaf660cdb6fec5f WatchSource:0}: Error finding container 2ae143e07b24e9b0cb1ed0b5cb463bef5b8b0f1aeae1b1118eaf660cdb6fec5f: Status 404 returned error can't find the container with id 2ae143e07b24e9b0cb1ed0b5cb463bef5b8b0f1aeae1b1118eaf660cdb6fec5f Mar 13 09:29:54 crc kubenswrapper[4930]: I0313 09:29:54.024107 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7f1370aa-666b-439e-9411-ec5e70360091-memberlist\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:54 crc kubenswrapper[4930]: E0313 09:29:54.024371 4930 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 13 09:29:54 crc kubenswrapper[4930]: E0313 09:29:54.024732 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/7f1370aa-666b-439e-9411-ec5e70360091-memberlist podName:7f1370aa-666b-439e-9411-ec5e70360091 nodeName:}" failed. No retries permitted until 2026-03-13 09:29:55.02469771 +0000 UTC m=+1035.774612427 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/7f1370aa-666b-439e-9411-ec5e70360091-memberlist") pod "speaker-hc6zf" (UID: "7f1370aa-666b-439e-9411-ec5e70360091") : secret "metallb-memberlist" not found Mar 13 09:29:54 crc kubenswrapper[4930]: I0313 09:29:54.047816 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 09:29:54 crc kubenswrapper[4930]: I0313 09:29:54.114205 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-9bhdc" event={"ID":"aa56d2a5-5b3a-487c-9248-21ef5689562f","Type":"ContainerStarted","Data":"54aac52da63da9f6e2134ae4af067569cf9545e3588d840ab16b2ee3eef8b684"} Mar 13 09:29:54 crc kubenswrapper[4930]: I0313 09:29:54.114266 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-9bhdc" event={"ID":"aa56d2a5-5b3a-487c-9248-21ef5689562f","Type":"ContainerStarted","Data":"2ae143e07b24e9b0cb1ed0b5cb463bef5b8b0f1aeae1b1118eaf660cdb6fec5f"} Mar 13 09:29:54 crc kubenswrapper[4930]: I0313 09:29:54.116299 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerStarted","Data":"3975156adf5b5d516b9f82c64440929cb85eeecd9478d49dc5b1d29ef4fbc208"} Mar 13 09:29:54 crc kubenswrapper[4930]: I0313 09:29:54.466639 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx"] Mar 13 09:29:55 crc kubenswrapper[4930]: I0313 09:29:55.039508 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7f1370aa-666b-439e-9411-ec5e70360091-memberlist\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:55 crc kubenswrapper[4930]: I0313 09:29:55.045362 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/7f1370aa-666b-439e-9411-ec5e70360091-memberlist\") pod \"speaker-hc6zf\" (UID: \"7f1370aa-666b-439e-9411-ec5e70360091\") " pod="metallb-system/speaker-hc6zf" Mar 13 09:29:55 crc kubenswrapper[4930]: I0313 09:29:55.064133 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hc6zf" Mar 13 09:29:55 crc kubenswrapper[4930]: W0313 09:29:55.094374 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7f1370aa_666b_439e_9411_ec5e70360091.slice/crio-54446f85fadf65108aa62c0536300bb78a05288b4e992ff2ce5f84691b1e0c88 WatchSource:0}: Error finding container 54446f85fadf65108aa62c0536300bb78a05288b4e992ff2ce5f84691b1e0c88: Status 404 returned error can't find the container with id 54446f85fadf65108aa62c0536300bb78a05288b4e992ff2ce5f84691b1e0c88 Mar 13 09:29:55 crc kubenswrapper[4930]: I0313 09:29:55.128121 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hc6zf" event={"ID":"7f1370aa-666b-439e-9411-ec5e70360091","Type":"ContainerStarted","Data":"54446f85fadf65108aa62c0536300bb78a05288b4e992ff2ce5f84691b1e0c88"} Mar 13 09:29:55 crc kubenswrapper[4930]: I0313 09:29:55.131603 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-9bhdc" event={"ID":"aa56d2a5-5b3a-487c-9248-21ef5689562f","Type":"ContainerStarted","Data":"7e1b9732a79a56f0a234932fc9a6aa3066181ee5c094baf5a2efa03f9b83eb90"} Mar 13 09:29:55 crc kubenswrapper[4930]: I0313 09:29:55.131725 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:29:55 crc kubenswrapper[4930]: I0313 09:29:55.144241 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" event={"ID":"bbf3e078-1f66-4bcc-a742-0835afcb3e79","Type":"ContainerStarted","Data":"7ce5a15d8d51fef92734fc00244a4953192bab0ff504a8d9f89dc85fcb4f4b1e"} Mar 13 09:29:55 crc kubenswrapper[4930]: I0313 09:29:55.155634 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-9bhdc" podStartSLOduration=2.155610726 podStartE2EDuration="2.155610726s" podCreationTimestamp="2026-03-13 09:29:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:29:55.15098234 +0000 UTC m=+1035.900897017" watchObservedRunningTime="2026-03-13 09:29:55.155610726 +0000 UTC m=+1035.905525403" Mar 13 09:29:56 crc kubenswrapper[4930]: I0313 09:29:56.154570 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hc6zf" event={"ID":"7f1370aa-666b-439e-9411-ec5e70360091","Type":"ContainerStarted","Data":"67f33a7e147933c6c9e941ecc2b3f615a8bb6bc4c2951961f905bd62fe4a7b72"} Mar 13 09:29:56 crc kubenswrapper[4930]: I0313 09:29:56.154624 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hc6zf" event={"ID":"7f1370aa-666b-439e-9411-ec5e70360091","Type":"ContainerStarted","Data":"b8fd9a8a2f4a99cfa5b81539c8e8c3d5bf5f01ce1422d2f280455d29d84073f6"} Mar 13 09:29:56 crc kubenswrapper[4930]: I0313 09:29:56.154752 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hc6zf" Mar 13 09:29:56 crc kubenswrapper[4930]: I0313 09:29:56.182224 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-hc6zf" podStartSLOduration=3.182202241 podStartE2EDuration="3.182202241s" podCreationTimestamp="2026-03-13 09:29:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:29:56.179932705 +0000 UTC m=+1036.929847382" watchObservedRunningTime="2026-03-13 09:29:56.182202241 +0000 UTC m=+1036.932116918" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.130444 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556570-qjp4p"] Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.131817 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556570-qjp4p" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.135052 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.135289 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.135478 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.143131 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49"] Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.144074 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556570-qjp4p"] Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.144156 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.146356 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.146590 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.149090 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49"] Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.227682 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rxvf\" (UniqueName: \"kubernetes.io/projected/25af5d21-ed16-4f4e-b49c-c78b965bc965-kube-api-access-8rxvf\") pod \"collect-profiles-29556570-gmd49\" (UID: \"25af5d21-ed16-4f4e-b49c-c78b965bc965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.227771 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdv4x\" (UniqueName: \"kubernetes.io/projected/1c44fc59-e3a9-448a-b593-d572109a37c4-kube-api-access-jdv4x\") pod \"auto-csr-approver-29556570-qjp4p\" (UID: \"1c44fc59-e3a9-448a-b593-d572109a37c4\") " pod="openshift-infra/auto-csr-approver-29556570-qjp4p" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.227833 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25af5d21-ed16-4f4e-b49c-c78b965bc965-secret-volume\") pod \"collect-profiles-29556570-gmd49\" (UID: \"25af5d21-ed16-4f4e-b49c-c78b965bc965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.227883 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25af5d21-ed16-4f4e-b49c-c78b965bc965-config-volume\") pod \"collect-profiles-29556570-gmd49\" (UID: \"25af5d21-ed16-4f4e-b49c-c78b965bc965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.329087 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25af5d21-ed16-4f4e-b49c-c78b965bc965-secret-volume\") pod \"collect-profiles-29556570-gmd49\" (UID: \"25af5d21-ed16-4f4e-b49c-c78b965bc965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.329173 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25af5d21-ed16-4f4e-b49c-c78b965bc965-config-volume\") pod \"collect-profiles-29556570-gmd49\" (UID: \"25af5d21-ed16-4f4e-b49c-c78b965bc965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.329205 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rxvf\" (UniqueName: \"kubernetes.io/projected/25af5d21-ed16-4f4e-b49c-c78b965bc965-kube-api-access-8rxvf\") pod \"collect-profiles-29556570-gmd49\" (UID: \"25af5d21-ed16-4f4e-b49c-c78b965bc965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.329276 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdv4x\" (UniqueName: \"kubernetes.io/projected/1c44fc59-e3a9-448a-b593-d572109a37c4-kube-api-access-jdv4x\") pod \"auto-csr-approver-29556570-qjp4p\" (UID: \"1c44fc59-e3a9-448a-b593-d572109a37c4\") " pod="openshift-infra/auto-csr-approver-29556570-qjp4p" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.330217 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25af5d21-ed16-4f4e-b49c-c78b965bc965-config-volume\") pod \"collect-profiles-29556570-gmd49\" (UID: \"25af5d21-ed16-4f4e-b49c-c78b965bc965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.345458 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdv4x\" (UniqueName: \"kubernetes.io/projected/1c44fc59-e3a9-448a-b593-d572109a37c4-kube-api-access-jdv4x\") pod \"auto-csr-approver-29556570-qjp4p\" (UID: \"1c44fc59-e3a9-448a-b593-d572109a37c4\") " pod="openshift-infra/auto-csr-approver-29556570-qjp4p" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.346380 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25af5d21-ed16-4f4e-b49c-c78b965bc965-secret-volume\") pod \"collect-profiles-29556570-gmd49\" (UID: \"25af5d21-ed16-4f4e-b49c-c78b965bc965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.348974 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rxvf\" (UniqueName: \"kubernetes.io/projected/25af5d21-ed16-4f4e-b49c-c78b965bc965-kube-api-access-8rxvf\") pod \"collect-profiles-29556570-gmd49\" (UID: \"25af5d21-ed16-4f4e-b49c-c78b965bc965\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.486564 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556570-qjp4p" Mar 13 09:30:00 crc kubenswrapper[4930]: I0313 09:30:00.492926 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:01 crc kubenswrapper[4930]: I0313 09:30:01.909330 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49"] Mar 13 09:30:01 crc kubenswrapper[4930]: W0313 09:30:01.923313 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25af5d21_ed16_4f4e_b49c_c78b965bc965.slice/crio-dcca1e2883266bee384998f5614733bbc737ccb1a86210caf3a9cdc73dbb9140 WatchSource:0}: Error finding container dcca1e2883266bee384998f5614733bbc737ccb1a86210caf3a9cdc73dbb9140: Status 404 returned error can't find the container with id dcca1e2883266bee384998f5614733bbc737ccb1a86210caf3a9cdc73dbb9140 Mar 13 09:30:01 crc kubenswrapper[4930]: I0313 09:30:01.981584 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 09:30:01 crc kubenswrapper[4930]: I0313 09:30:01.992349 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556570-qjp4p"] Mar 13 09:30:02 crc kubenswrapper[4930]: I0313 09:30:02.205897 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556570-qjp4p" event={"ID":"1c44fc59-e3a9-448a-b593-d572109a37c4","Type":"ContainerStarted","Data":"b684997e4bdec3817e5df9b59a44a2dc93533eb4a951264faf969e47caef6c18"} Mar 13 09:30:02 crc kubenswrapper[4930]: I0313 09:30:02.207387 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" event={"ID":"25af5d21-ed16-4f4e-b49c-c78b965bc965","Type":"ContainerStarted","Data":"c3f6cdc3f893b1ef270098459978b42705f2b9a5459595812e1a5430879a177b"} Mar 13 09:30:02 crc kubenswrapper[4930]: I0313 09:30:02.207420 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" event={"ID":"25af5d21-ed16-4f4e-b49c-c78b965bc965","Type":"ContainerStarted","Data":"dcca1e2883266bee384998f5614733bbc737ccb1a86210caf3a9cdc73dbb9140"} Mar 13 09:30:02 crc kubenswrapper[4930]: I0313 09:30:02.210142 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" event={"ID":"bbf3e078-1f66-4bcc-a742-0835afcb3e79","Type":"ContainerStarted","Data":"9fe44cdc98892fa6b9a768d0a23c5444c75a778e1a00875feae8c0c303a524ed"} Mar 13 09:30:02 crc kubenswrapper[4930]: I0313 09:30:02.210347 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 09:30:02 crc kubenswrapper[4930]: I0313 09:30:02.212374 4930 generic.go:334] "Generic (PLEG): container finished" podID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerID="e40b27199e8bfa30745304d7e2d8a5bf2b87c25517a295d0b17f10f052a7011f" exitCode=0 Mar 13 09:30:02 crc kubenswrapper[4930]: I0313 09:30:02.212414 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerDied","Data":"e40b27199e8bfa30745304d7e2d8a5bf2b87c25517a295d0b17f10f052a7011f"} Mar 13 09:30:02 crc kubenswrapper[4930]: I0313 09:30:02.244006 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" podStartSLOduration=2.243980075 podStartE2EDuration="2.243980075s" podCreationTimestamp="2026-03-13 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:30:02.224378457 +0000 UTC m=+1042.974293164" watchObservedRunningTime="2026-03-13 09:30:02.243980075 +0000 UTC m=+1042.993894792" Mar 13 09:30:02 crc kubenswrapper[4930]: I0313 09:30:02.254592 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" podStartSLOduration=2.198640578 podStartE2EDuration="9.254566229s" podCreationTimestamp="2026-03-13 09:29:53 +0000 UTC" firstStartedPulling="2026-03-13 09:29:54.475608708 +0000 UTC m=+1035.225523385" lastFinishedPulling="2026-03-13 09:30:01.531534359 +0000 UTC m=+1042.281449036" observedRunningTime="2026-03-13 09:30:02.248322694 +0000 UTC m=+1042.998237371" watchObservedRunningTime="2026-03-13 09:30:02.254566229 +0000 UTC m=+1043.004480956" Mar 13 09:30:03 crc kubenswrapper[4930]: I0313 09:30:03.220699 4930 generic.go:334] "Generic (PLEG): container finished" podID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerID="5fdfb214fe51a544e2b365ac8222632aa7d43ab323ec1ad2ca7c9abe1c9c919e" exitCode=0 Mar 13 09:30:03 crc kubenswrapper[4930]: I0313 09:30:03.221046 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerDied","Data":"5fdfb214fe51a544e2b365ac8222632aa7d43ab323ec1ad2ca7c9abe1c9c919e"} Mar 13 09:30:03 crc kubenswrapper[4930]: I0313 09:30:03.223122 4930 generic.go:334] "Generic (PLEG): container finished" podID="25af5d21-ed16-4f4e-b49c-c78b965bc965" containerID="c3f6cdc3f893b1ef270098459978b42705f2b9a5459595812e1a5430879a177b" exitCode=0 Mar 13 09:30:03 crc kubenswrapper[4930]: I0313 09:30:03.223979 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" event={"ID":"25af5d21-ed16-4f4e-b49c-c78b965bc965","Type":"ContainerDied","Data":"c3f6cdc3f893b1ef270098459978b42705f2b9a5459595812e1a5430879a177b"} Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.236081 4930 generic.go:334] "Generic (PLEG): container finished" podID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerID="9f22a67eebc8f54d5280f3cd64b19ba231be6c8cf1e5ef9d84900b5751aa642a" exitCode=0 Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.236268 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerDied","Data":"9f22a67eebc8f54d5280f3cd64b19ba231be6c8cf1e5ef9d84900b5751aa642a"} Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.238832 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556570-qjp4p" event={"ID":"1c44fc59-e3a9-448a-b593-d572109a37c4","Type":"ContainerStarted","Data":"ede3f1783e9d36de90ca201f79679248e68b7f2d10a2509424618bd62ed0b6fa"} Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.286775 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556570-qjp4p" podStartSLOduration=2.32078758 podStartE2EDuration="4.286747937s" podCreationTimestamp="2026-03-13 09:30:00 +0000 UTC" firstStartedPulling="2026-03-13 09:30:01.981225347 +0000 UTC m=+1042.731140024" lastFinishedPulling="2026-03-13 09:30:03.947185704 +0000 UTC m=+1044.697100381" observedRunningTime="2026-03-13 09:30:04.276793779 +0000 UTC m=+1045.026708456" watchObservedRunningTime="2026-03-13 09:30:04.286747937 +0000 UTC m=+1045.036662624" Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.550827 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.616049 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25af5d21-ed16-4f4e-b49c-c78b965bc965-config-volume\") pod \"25af5d21-ed16-4f4e-b49c-c78b965bc965\" (UID: \"25af5d21-ed16-4f4e-b49c-c78b965bc965\") " Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.616164 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rxvf\" (UniqueName: \"kubernetes.io/projected/25af5d21-ed16-4f4e-b49c-c78b965bc965-kube-api-access-8rxvf\") pod \"25af5d21-ed16-4f4e-b49c-c78b965bc965\" (UID: \"25af5d21-ed16-4f4e-b49c-c78b965bc965\") " Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.616218 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25af5d21-ed16-4f4e-b49c-c78b965bc965-secret-volume\") pod \"25af5d21-ed16-4f4e-b49c-c78b965bc965\" (UID: \"25af5d21-ed16-4f4e-b49c-c78b965bc965\") " Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.618558 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25af5d21-ed16-4f4e-b49c-c78b965bc965-config-volume" (OuterVolumeSpecName: "config-volume") pod "25af5d21-ed16-4f4e-b49c-c78b965bc965" (UID: "25af5d21-ed16-4f4e-b49c-c78b965bc965"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.624686 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25af5d21-ed16-4f4e-b49c-c78b965bc965-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "25af5d21-ed16-4f4e-b49c-c78b965bc965" (UID: "25af5d21-ed16-4f4e-b49c-c78b965bc965"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.624909 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25af5d21-ed16-4f4e-b49c-c78b965bc965-kube-api-access-8rxvf" (OuterVolumeSpecName: "kube-api-access-8rxvf") pod "25af5d21-ed16-4f4e-b49c-c78b965bc965" (UID: "25af5d21-ed16-4f4e-b49c-c78b965bc965"). InnerVolumeSpecName "kube-api-access-8rxvf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.718074 4930 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/25af5d21-ed16-4f4e-b49c-c78b965bc965-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.718109 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rxvf\" (UniqueName: \"kubernetes.io/projected/25af5d21-ed16-4f4e-b49c-c78b965bc965-kube-api-access-8rxvf\") on node \"crc\" DevicePath \"\"" Mar 13 09:30:04 crc kubenswrapper[4930]: I0313 09:30:04.718120 4930 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/25af5d21-ed16-4f4e-b49c-c78b965bc965-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 09:30:05 crc kubenswrapper[4930]: I0313 09:30:05.070833 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-hc6zf" Mar 13 09:30:05 crc kubenswrapper[4930]: I0313 09:30:05.257328 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" event={"ID":"25af5d21-ed16-4f4e-b49c-c78b965bc965","Type":"ContainerDied","Data":"dcca1e2883266bee384998f5614733bbc737ccb1a86210caf3a9cdc73dbb9140"} Mar 13 09:30:05 crc kubenswrapper[4930]: I0313 09:30:05.257390 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcca1e2883266bee384998f5614733bbc737ccb1a86210caf3a9cdc73dbb9140" Mar 13 09:30:05 crc kubenswrapper[4930]: I0313 09:30:05.257341 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49" Mar 13 09:30:05 crc kubenswrapper[4930]: I0313 09:30:05.270539 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerStarted","Data":"d067aa2d7b2d2cfa40c2499adb57020b1d284dbe7733d78e99a8a0e064a5e900"} Mar 13 09:30:05 crc kubenswrapper[4930]: I0313 09:30:05.270591 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerStarted","Data":"b4d46e8e6ad912142d13c6ca262c95f78dc8d0482e5645e4c90f87e0ae5bb71c"} Mar 13 09:30:05 crc kubenswrapper[4930]: I0313 09:30:05.270605 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerStarted","Data":"cf928213d0e82cb32fb7a79a685d5ec07cf8fc0c655c7f1144ef7462392dc88e"} Mar 13 09:30:05 crc kubenswrapper[4930]: I0313 09:30:05.270620 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerStarted","Data":"da0e90531521761cb303522a5a41ba4b7c379d88504e0716cbb1fa3ec99fdc76"} Mar 13 09:30:05 crc kubenswrapper[4930]: I0313 09:30:05.270633 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerStarted","Data":"0b385e60f422e66ee531f65506a5ebd87f2b5f04b2d87a40d22a6bd2e3201027"} Mar 13 09:30:05 crc kubenswrapper[4930]: I0313 09:30:05.283029 4930 generic.go:334] "Generic (PLEG): container finished" podID="1c44fc59-e3a9-448a-b593-d572109a37c4" containerID="ede3f1783e9d36de90ca201f79679248e68b7f2d10a2509424618bd62ed0b6fa" exitCode=0 Mar 13 09:30:05 crc kubenswrapper[4930]: I0313 09:30:05.283078 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556570-qjp4p" event={"ID":"1c44fc59-e3a9-448a-b593-d572109a37c4","Type":"ContainerDied","Data":"ede3f1783e9d36de90ca201f79679248e68b7f2d10a2509424618bd62ed0b6fa"} Mar 13 09:30:06 crc kubenswrapper[4930]: I0313 09:30:06.293915 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerStarted","Data":"72d9c4437520d4232c07492bc4c8ea10517c50e80165e663a6fd689d5d92fa89"} Mar 13 09:30:06 crc kubenswrapper[4930]: I0313 09:30:06.325184 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-krmnx" podStartSLOduration=5.490738215 podStartE2EDuration="13.325164389s" podCreationTimestamp="2026-03-13 09:29:53 +0000 UTC" firstStartedPulling="2026-03-13 09:29:53.673800315 +0000 UTC m=+1034.423714992" lastFinishedPulling="2026-03-13 09:30:01.508226499 +0000 UTC m=+1042.258141166" observedRunningTime="2026-03-13 09:30:06.315625351 +0000 UTC m=+1047.065540048" watchObservedRunningTime="2026-03-13 09:30:06.325164389 +0000 UTC m=+1047.075079066" Mar 13 09:30:06 crc kubenswrapper[4930]: I0313 09:30:06.608233 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556570-qjp4p" Mar 13 09:30:06 crc kubenswrapper[4930]: I0313 09:30:06.659701 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdv4x\" (UniqueName: \"kubernetes.io/projected/1c44fc59-e3a9-448a-b593-d572109a37c4-kube-api-access-jdv4x\") pod \"1c44fc59-e3a9-448a-b593-d572109a37c4\" (UID: \"1c44fc59-e3a9-448a-b593-d572109a37c4\") " Mar 13 09:30:06 crc kubenswrapper[4930]: I0313 09:30:06.665806 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c44fc59-e3a9-448a-b593-d572109a37c4-kube-api-access-jdv4x" (OuterVolumeSpecName: "kube-api-access-jdv4x") pod "1c44fc59-e3a9-448a-b593-d572109a37c4" (UID: "1c44fc59-e3a9-448a-b593-d572109a37c4"). InnerVolumeSpecName "kube-api-access-jdv4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:30:06 crc kubenswrapper[4930]: I0313 09:30:06.762008 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdv4x\" (UniqueName: \"kubernetes.io/projected/1c44fc59-e3a9-448a-b593-d572109a37c4-kube-api-access-jdv4x\") on node \"crc\" DevicePath \"\"" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.306482 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556570-qjp4p" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.306525 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556570-qjp4p" event={"ID":"1c44fc59-e3a9-448a-b593-d572109a37c4","Type":"ContainerDied","Data":"b684997e4bdec3817e5df9b59a44a2dc93533eb4a951264faf969e47caef6c18"} Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.307042 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b684997e4bdec3817e5df9b59a44a2dc93533eb4a951264faf969e47caef6c18" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.307065 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-krmnx" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.361116 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556564-rw6rl"] Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.370337 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556564-rw6rl"] Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.774911 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-qxhwt"] Mar 13 09:30:07 crc kubenswrapper[4930]: E0313 09:30:07.775291 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c44fc59-e3a9-448a-b593-d572109a37c4" containerName="oc" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.775313 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c44fc59-e3a9-448a-b593-d572109a37c4" containerName="oc" Mar 13 09:30:07 crc kubenswrapper[4930]: E0313 09:30:07.775348 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25af5d21-ed16-4f4e-b49c-c78b965bc965" containerName="collect-profiles" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.775356 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="25af5d21-ed16-4f4e-b49c-c78b965bc965" containerName="collect-profiles" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.775565 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c44fc59-e3a9-448a-b593-d572109a37c4" containerName="oc" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.775596 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="25af5d21-ed16-4f4e-b49c-c78b965bc965" containerName="collect-profiles" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.776254 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qxhwt" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.779389 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.779541 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-xv5cf" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.785133 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.792132 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qxhwt"] Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.877453 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsjps\" (UniqueName: \"kubernetes.io/projected/36d90cf0-9e7b-4283-a395-3fd06725d911-kube-api-access-wsjps\") pod \"openstack-operator-index-qxhwt\" (UID: \"36d90cf0-9e7b-4283-a395-3fd06725d911\") " pod="openstack-operators/openstack-operator-index-qxhwt" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.984064 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsjps\" (UniqueName: \"kubernetes.io/projected/36d90cf0-9e7b-4283-a395-3fd06725d911-kube-api-access-wsjps\") pod \"openstack-operator-index-qxhwt\" (UID: \"36d90cf0-9e7b-4283-a395-3fd06725d911\") " pod="openstack-operators/openstack-operator-index-qxhwt" Mar 13 09:30:07 crc kubenswrapper[4930]: I0313 09:30:07.994162 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9dc22f8-83b4-4579-80b2-5b8d095acc4b" path="/var/lib/kubelet/pods/b9dc22f8-83b4-4579-80b2-5b8d095acc4b/volumes" Mar 13 09:30:08 crc kubenswrapper[4930]: I0313 09:30:08.011573 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsjps\" (UniqueName: \"kubernetes.io/projected/36d90cf0-9e7b-4283-a395-3fd06725d911-kube-api-access-wsjps\") pod \"openstack-operator-index-qxhwt\" (UID: \"36d90cf0-9e7b-4283-a395-3fd06725d911\") " pod="openstack-operators/openstack-operator-index-qxhwt" Mar 13 09:30:08 crc kubenswrapper[4930]: I0313 09:30:08.092018 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qxhwt" Mar 13 09:30:08 crc kubenswrapper[4930]: I0313 09:30:08.480759 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-krmnx" Mar 13 09:30:08 crc kubenswrapper[4930]: I0313 09:30:08.522885 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-krmnx" Mar 13 09:30:08 crc kubenswrapper[4930]: I0313 09:30:08.569630 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-qxhwt"] Mar 13 09:30:08 crc kubenswrapper[4930]: W0313 09:30:08.574030 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod36d90cf0_9e7b_4283_a395_3fd06725d911.slice/crio-d29a4c6bfe9696794466a0800795d2765c8279649cc9072fd20ce6fd7febde5b WatchSource:0}: Error finding container d29a4c6bfe9696794466a0800795d2765c8279649cc9072fd20ce6fd7febde5b: Status 404 returned error can't find the container with id d29a4c6bfe9696794466a0800795d2765c8279649cc9072fd20ce6fd7febde5b Mar 13 09:30:09 crc kubenswrapper[4930]: I0313 09:30:09.322445 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qxhwt" event={"ID":"36d90cf0-9e7b-4283-a395-3fd06725d911","Type":"ContainerStarted","Data":"d29a4c6bfe9696794466a0800795d2765c8279649cc9072fd20ce6fd7febde5b"} Mar 13 09:30:11 crc kubenswrapper[4930]: I0313 09:30:11.154616 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-qxhwt"] Mar 13 09:30:11 crc kubenswrapper[4930]: I0313 09:30:11.769763 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-5rhcc"] Mar 13 09:30:11 crc kubenswrapper[4930]: I0313 09:30:11.771482 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 09:30:11 crc kubenswrapper[4930]: I0313 09:30:11.780490 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5rhcc"] Mar 13 09:30:11 crc kubenswrapper[4930]: I0313 09:30:11.921338 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5c2n\" (UniqueName: \"kubernetes.io/projected/08a4c763-07ef-4163-84ea-7a9bb39de0f1-kube-api-access-w5c2n\") pod \"openstack-operator-index-5rhcc\" (UID: \"08a4c763-07ef-4163-84ea-7a9bb39de0f1\") " pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 09:30:12 crc kubenswrapper[4930]: I0313 09:30:12.022972 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5c2n\" (UniqueName: \"kubernetes.io/projected/08a4c763-07ef-4163-84ea-7a9bb39de0f1-kube-api-access-w5c2n\") pod \"openstack-operator-index-5rhcc\" (UID: \"08a4c763-07ef-4163-84ea-7a9bb39de0f1\") " pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 09:30:12 crc kubenswrapper[4930]: I0313 09:30:12.042973 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5c2n\" (UniqueName: \"kubernetes.io/projected/08a4c763-07ef-4163-84ea-7a9bb39de0f1-kube-api-access-w5c2n\") pod \"openstack-operator-index-5rhcc\" (UID: \"08a4c763-07ef-4163-84ea-7a9bb39de0f1\") " pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 09:30:12 crc kubenswrapper[4930]: I0313 09:30:12.144797 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 09:30:12 crc kubenswrapper[4930]: I0313 09:30:12.946002 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-5rhcc"] Mar 13 09:30:13 crc kubenswrapper[4930]: W0313 09:30:13.414158 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod08a4c763_07ef_4163_84ea_7a9bb39de0f1.slice/crio-6d5401bb5cc8545d1f100ea17213e0dda635fd3396b5cf0ff1454cd6d3c18f4e WatchSource:0}: Error finding container 6d5401bb5cc8545d1f100ea17213e0dda635fd3396b5cf0ff1454cd6d3c18f4e: Status 404 returned error can't find the container with id 6d5401bb5cc8545d1f100ea17213e0dda635fd3396b5cf0ff1454cd6d3c18f4e Mar 13 09:30:13 crc kubenswrapper[4930]: I0313 09:30:13.559468 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 09:30:14 crc kubenswrapper[4930]: I0313 09:30:14.055540 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 09:30:14 crc kubenswrapper[4930]: I0313 09:30:14.362139 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5rhcc" event={"ID":"08a4c763-07ef-4163-84ea-7a9bb39de0f1","Type":"ContainerStarted","Data":"e1607208fa9fe34d7a2cd486255d55261b51f641995f51b7b3fd83a2ff83ad9c"} Mar 13 09:30:14 crc kubenswrapper[4930]: I0313 09:30:14.362900 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5rhcc" event={"ID":"08a4c763-07ef-4163-84ea-7a9bb39de0f1","Type":"ContainerStarted","Data":"6d5401bb5cc8545d1f100ea17213e0dda635fd3396b5cf0ff1454cd6d3c18f4e"} Mar 13 09:30:14 crc kubenswrapper[4930]: I0313 09:30:14.364170 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qxhwt" event={"ID":"36d90cf0-9e7b-4283-a395-3fd06725d911","Type":"ContainerStarted","Data":"b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a"} Mar 13 09:30:14 crc kubenswrapper[4930]: I0313 09:30:14.364234 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-qxhwt" podUID="36d90cf0-9e7b-4283-a395-3fd06725d911" containerName="registry-server" containerID="cri-o://b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a" gracePeriod=2 Mar 13 09:30:14 crc kubenswrapper[4930]: I0313 09:30:14.377903 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-5rhcc" podStartSLOduration=3.334641935 podStartE2EDuration="3.377888703s" podCreationTimestamp="2026-03-13 09:30:11 +0000 UTC" firstStartedPulling="2026-03-13 09:30:13.422089462 +0000 UTC m=+1054.172004139" lastFinishedPulling="2026-03-13 09:30:13.46533623 +0000 UTC m=+1054.215250907" observedRunningTime="2026-03-13 09:30:14.376254122 +0000 UTC m=+1055.126168799" watchObservedRunningTime="2026-03-13 09:30:14.377888703 +0000 UTC m=+1055.127803380" Mar 13 09:30:14 crc kubenswrapper[4930]: I0313 09:30:14.397696 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-qxhwt" podStartSLOduration=2.510140786 podStartE2EDuration="7.397675096s" podCreationTimestamp="2026-03-13 09:30:07 +0000 UTC" firstStartedPulling="2026-03-13 09:30:08.576497317 +0000 UTC m=+1049.326412034" lastFinishedPulling="2026-03-13 09:30:13.464031677 +0000 UTC m=+1054.213946344" observedRunningTime="2026-03-13 09:30:14.390120668 +0000 UTC m=+1055.140035345" watchObservedRunningTime="2026-03-13 09:30:14.397675096 +0000 UTC m=+1055.147589773" Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.260899 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qxhwt" Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.375258 4930 generic.go:334] "Generic (PLEG): container finished" podID="36d90cf0-9e7b-4283-a395-3fd06725d911" containerID="b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a" exitCode=0 Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.375317 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-qxhwt" Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.375361 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qxhwt" event={"ID":"36d90cf0-9e7b-4283-a395-3fd06725d911","Type":"ContainerDied","Data":"b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a"} Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.375387 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-qxhwt" event={"ID":"36d90cf0-9e7b-4283-a395-3fd06725d911","Type":"ContainerDied","Data":"d29a4c6bfe9696794466a0800795d2765c8279649cc9072fd20ce6fd7febde5b"} Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.375406 4930 scope.go:117] "RemoveContainer" containerID="b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a" Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.383540 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsjps\" (UniqueName: \"kubernetes.io/projected/36d90cf0-9e7b-4283-a395-3fd06725d911-kube-api-access-wsjps\") pod \"36d90cf0-9e7b-4283-a395-3fd06725d911\" (UID: \"36d90cf0-9e7b-4283-a395-3fd06725d911\") " Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.390594 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36d90cf0-9e7b-4283-a395-3fd06725d911-kube-api-access-wsjps" (OuterVolumeSpecName: "kube-api-access-wsjps") pod "36d90cf0-9e7b-4283-a395-3fd06725d911" (UID: "36d90cf0-9e7b-4283-a395-3fd06725d911"). InnerVolumeSpecName "kube-api-access-wsjps". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.400450 4930 scope.go:117] "RemoveContainer" containerID="b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a" Mar 13 09:30:15 crc kubenswrapper[4930]: E0313 09:30:15.400921 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a\": container with ID starting with b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a not found: ID does not exist" containerID="b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a" Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.400958 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a"} err="failed to get container status \"b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a\": rpc error: code = NotFound desc = could not find container \"b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a\": container with ID starting with b2e184921e019a9502ff33e1bc01cb7a7d309f99a22fefa9961d7400ea648a2a not found: ID does not exist" Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.485359 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsjps\" (UniqueName: \"kubernetes.io/projected/36d90cf0-9e7b-4283-a395-3fd06725d911-kube-api-access-wsjps\") on node \"crc\" DevicePath \"\"" Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.706133 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-qxhwt"] Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.713798 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-qxhwt"] Mar 13 09:30:15 crc kubenswrapper[4930]: I0313 09:30:15.979459 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36d90cf0-9e7b-4283-a395-3fd06725d911" path="/var/lib/kubelet/pods/36d90cf0-9e7b-4283-a395-3fd06725d911/volumes" Mar 13 09:30:22 crc kubenswrapper[4930]: I0313 09:30:22.145476 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 09:30:22 crc kubenswrapper[4930]: I0313 09:30:22.145964 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 09:30:22 crc kubenswrapper[4930]: I0313 09:30:22.177175 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 09:30:22 crc kubenswrapper[4930]: I0313 09:30:22.455632 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.193540 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk"] Mar 13 09:30:23 crc kubenswrapper[4930]: E0313 09:30:23.193831 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d90cf0-9e7b-4283-a395-3fd06725d911" containerName="registry-server" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.193845 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d90cf0-9e7b-4283-a395-3fd06725d911" containerName="registry-server" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.194013 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="36d90cf0-9e7b-4283-a395-3fd06725d911" containerName="registry-server" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.195081 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.196620 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-n6wnq" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.204635 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk"] Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.313950 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j262z\" (UniqueName: \"kubernetes.io/projected/bb65fee3-b414-4020-8669-4c8ee27ca0ef-kube-api-access-j262z\") pod \"aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk\" (UID: \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\") " pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.314007 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb65fee3-b414-4020-8669-4c8ee27ca0ef-util\") pod \"aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk\" (UID: \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\") " pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.314040 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb65fee3-b414-4020-8669-4c8ee27ca0ef-bundle\") pod \"aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk\" (UID: \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\") " pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.415735 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j262z\" (UniqueName: \"kubernetes.io/projected/bb65fee3-b414-4020-8669-4c8ee27ca0ef-kube-api-access-j262z\") pod \"aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk\" (UID: \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\") " pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.415842 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb65fee3-b414-4020-8669-4c8ee27ca0ef-util\") pod \"aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk\" (UID: \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\") " pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.415903 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb65fee3-b414-4020-8669-4c8ee27ca0ef-bundle\") pod \"aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk\" (UID: \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\") " pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.416351 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb65fee3-b414-4020-8669-4c8ee27ca0ef-util\") pod \"aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk\" (UID: \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\") " pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.416808 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb65fee3-b414-4020-8669-4c8ee27ca0ef-bundle\") pod \"aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk\" (UID: \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\") " pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.446761 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j262z\" (UniqueName: \"kubernetes.io/projected/bb65fee3-b414-4020-8669-4c8ee27ca0ef-kube-api-access-j262z\") pod \"aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk\" (UID: \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\") " pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.485829 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-krmnx" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.511339 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:23 crc kubenswrapper[4930]: I0313 09:30:23.926503 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk"] Mar 13 09:30:23 crc kubenswrapper[4930]: W0313 09:30:23.927543 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb65fee3_b414_4020_8669_4c8ee27ca0ef.slice/crio-e1c21a2c9e2c7eb31ca656ac4c7b1d72a4da990e29efa205777777f9157417c1 WatchSource:0}: Error finding container e1c21a2c9e2c7eb31ca656ac4c7b1d72a4da990e29efa205777777f9157417c1: Status 404 returned error can't find the container with id e1c21a2c9e2c7eb31ca656ac4c7b1d72a4da990e29efa205777777f9157417c1 Mar 13 09:30:24 crc kubenswrapper[4930]: I0313 09:30:24.451868 4930 generic.go:334] "Generic (PLEG): container finished" podID="bb65fee3-b414-4020-8669-4c8ee27ca0ef" containerID="c67381cb2313ad3feb518ee759c3ffe00d9a348893e1bf981e5df7c58bb278be" exitCode=0 Mar 13 09:30:24 crc kubenswrapper[4930]: I0313 09:30:24.452254 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" event={"ID":"bb65fee3-b414-4020-8669-4c8ee27ca0ef","Type":"ContainerDied","Data":"c67381cb2313ad3feb518ee759c3ffe00d9a348893e1bf981e5df7c58bb278be"} Mar 13 09:30:24 crc kubenswrapper[4930]: I0313 09:30:24.452305 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" event={"ID":"bb65fee3-b414-4020-8669-4c8ee27ca0ef","Type":"ContainerStarted","Data":"e1c21a2c9e2c7eb31ca656ac4c7b1d72a4da990e29efa205777777f9157417c1"} Mar 13 09:30:25 crc kubenswrapper[4930]: I0313 09:30:25.462803 4930 generic.go:334] "Generic (PLEG): container finished" podID="bb65fee3-b414-4020-8669-4c8ee27ca0ef" containerID="995fd6eb656e58d716eba8b6df7532611f391009b925b747d7b5f87da56b8ebe" exitCode=0 Mar 13 09:30:25 crc kubenswrapper[4930]: I0313 09:30:25.462839 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" event={"ID":"bb65fee3-b414-4020-8669-4c8ee27ca0ef","Type":"ContainerDied","Data":"995fd6eb656e58d716eba8b6df7532611f391009b925b747d7b5f87da56b8ebe"} Mar 13 09:30:26 crc kubenswrapper[4930]: I0313 09:30:26.477853 4930 generic.go:334] "Generic (PLEG): container finished" podID="bb65fee3-b414-4020-8669-4c8ee27ca0ef" containerID="fb0dcccd4014aeaa8ef68e5bf871063785eab0298317b2313b37d1461b89f601" exitCode=0 Mar 13 09:30:26 crc kubenswrapper[4930]: I0313 09:30:26.477943 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" event={"ID":"bb65fee3-b414-4020-8669-4c8ee27ca0ef","Type":"ContainerDied","Data":"fb0dcccd4014aeaa8ef68e5bf871063785eab0298317b2313b37d1461b89f601"} Mar 13 09:30:27 crc kubenswrapper[4930]: I0313 09:30:27.894783 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:27 crc kubenswrapper[4930]: I0313 09:30:27.987125 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb65fee3-b414-4020-8669-4c8ee27ca0ef-util\") pod \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\" (UID: \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\") " Mar 13 09:30:27 crc kubenswrapper[4930]: I0313 09:30:27.987203 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb65fee3-b414-4020-8669-4c8ee27ca0ef-bundle\") pod \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\" (UID: \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\") " Mar 13 09:30:27 crc kubenswrapper[4930]: I0313 09:30:27.987357 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j262z\" (UniqueName: \"kubernetes.io/projected/bb65fee3-b414-4020-8669-4c8ee27ca0ef-kube-api-access-j262z\") pod \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\" (UID: \"bb65fee3-b414-4020-8669-4c8ee27ca0ef\") " Mar 13 09:30:27 crc kubenswrapper[4930]: I0313 09:30:27.987994 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb65fee3-b414-4020-8669-4c8ee27ca0ef-bundle" (OuterVolumeSpecName: "bundle") pod "bb65fee3-b414-4020-8669-4c8ee27ca0ef" (UID: "bb65fee3-b414-4020-8669-4c8ee27ca0ef"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:30:27 crc kubenswrapper[4930]: I0313 09:30:27.988778 4930 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb65fee3-b414-4020-8669-4c8ee27ca0ef-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:30:27 crc kubenswrapper[4930]: I0313 09:30:27.992244 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb65fee3-b414-4020-8669-4c8ee27ca0ef-kube-api-access-j262z" (OuterVolumeSpecName: "kube-api-access-j262z") pod "bb65fee3-b414-4020-8669-4c8ee27ca0ef" (UID: "bb65fee3-b414-4020-8669-4c8ee27ca0ef"). InnerVolumeSpecName "kube-api-access-j262z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:30:28 crc kubenswrapper[4930]: I0313 09:30:28.001694 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb65fee3-b414-4020-8669-4c8ee27ca0ef-util" (OuterVolumeSpecName: "util") pod "bb65fee3-b414-4020-8669-4c8ee27ca0ef" (UID: "bb65fee3-b414-4020-8669-4c8ee27ca0ef"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:30:28 crc kubenswrapper[4930]: I0313 09:30:28.090650 4930 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb65fee3-b414-4020-8669-4c8ee27ca0ef-util\") on node \"crc\" DevicePath \"\"" Mar 13 09:30:28 crc kubenswrapper[4930]: I0313 09:30:28.090677 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j262z\" (UniqueName: \"kubernetes.io/projected/bb65fee3-b414-4020-8669-4c8ee27ca0ef-kube-api-access-j262z\") on node \"crc\" DevicePath \"\"" Mar 13 09:30:28 crc kubenswrapper[4930]: I0313 09:30:28.504989 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" event={"ID":"bb65fee3-b414-4020-8669-4c8ee27ca0ef","Type":"ContainerDied","Data":"e1c21a2c9e2c7eb31ca656ac4c7b1d72a4da990e29efa205777777f9157417c1"} Mar 13 09:30:28 crc kubenswrapper[4930]: I0313 09:30:28.505047 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1c21a2c9e2c7eb31ca656ac4c7b1d72a4da990e29efa205777777f9157417c1" Mar 13 09:30:28 crc kubenswrapper[4930]: I0313 09:30:28.505162 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/aefe4be7d9dfb952bcef84ae750a582140c77a17d6a336ca6408545236wbhtk" Mar 13 09:30:29 crc kubenswrapper[4930]: I0313 09:30:29.984455 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj"] Mar 13 09:30:29 crc kubenswrapper[4930]: E0313 09:30:29.985021 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb65fee3-b414-4020-8669-4c8ee27ca0ef" containerName="extract" Mar 13 09:30:29 crc kubenswrapper[4930]: I0313 09:30:29.985036 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb65fee3-b414-4020-8669-4c8ee27ca0ef" containerName="extract" Mar 13 09:30:29 crc kubenswrapper[4930]: E0313 09:30:29.985058 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb65fee3-b414-4020-8669-4c8ee27ca0ef" containerName="pull" Mar 13 09:30:29 crc kubenswrapper[4930]: I0313 09:30:29.985065 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb65fee3-b414-4020-8669-4c8ee27ca0ef" containerName="pull" Mar 13 09:30:29 crc kubenswrapper[4930]: E0313 09:30:29.985087 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb65fee3-b414-4020-8669-4c8ee27ca0ef" containerName="util" Mar 13 09:30:29 crc kubenswrapper[4930]: I0313 09:30:29.985094 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb65fee3-b414-4020-8669-4c8ee27ca0ef" containerName="util" Mar 13 09:30:29 crc kubenswrapper[4930]: I0313 09:30:29.985254 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb65fee3-b414-4020-8669-4c8ee27ca0ef" containerName="extract" Mar 13 09:30:29 crc kubenswrapper[4930]: I0313 09:30:29.985979 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" Mar 13 09:30:29 crc kubenswrapper[4930]: I0313 09:30:29.993011 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-89lsd" Mar 13 09:30:30 crc kubenswrapper[4930]: I0313 09:30:30.012941 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj"] Mar 13 09:30:30 crc kubenswrapper[4930]: I0313 09:30:30.125079 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hqpst\" (UniqueName: \"kubernetes.io/projected/404966af-7f68-4667-b535-81136694bdb9-kube-api-access-hqpst\") pod \"openstack-operator-controller-init-ff65d8fd4-h6zgj\" (UID: \"404966af-7f68-4667-b535-81136694bdb9\") " pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" Mar 13 09:30:30 crc kubenswrapper[4930]: I0313 09:30:30.226647 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hqpst\" (UniqueName: \"kubernetes.io/projected/404966af-7f68-4667-b535-81136694bdb9-kube-api-access-hqpst\") pod \"openstack-operator-controller-init-ff65d8fd4-h6zgj\" (UID: \"404966af-7f68-4667-b535-81136694bdb9\") " pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" Mar 13 09:30:30 crc kubenswrapper[4930]: I0313 09:30:30.247657 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hqpst\" (UniqueName: \"kubernetes.io/projected/404966af-7f68-4667-b535-81136694bdb9-kube-api-access-hqpst\") pod \"openstack-operator-controller-init-ff65d8fd4-h6zgj\" (UID: \"404966af-7f68-4667-b535-81136694bdb9\") " pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" Mar 13 09:30:30 crc kubenswrapper[4930]: I0313 09:30:30.313739 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" Mar 13 09:30:30 crc kubenswrapper[4930]: I0313 09:30:30.899147 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj"] Mar 13 09:30:31 crc kubenswrapper[4930]: I0313 09:30:31.533017 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" event={"ID":"404966af-7f68-4667-b535-81136694bdb9","Type":"ContainerStarted","Data":"2b66db619af19c56e6caadc72c95b4ad2f96f6fb4947ca6f6e6f474db1f76d83"} Mar 13 09:30:35 crc kubenswrapper[4930]: I0313 09:30:35.563781 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" event={"ID":"404966af-7f68-4667-b535-81136694bdb9","Type":"ContainerStarted","Data":"a91a2735163b8e5e2a6713ead07555eebc20e9f99df7a57d4fca28583cca41ed"} Mar 13 09:30:35 crc kubenswrapper[4930]: I0313 09:30:35.564471 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" Mar 13 09:30:35 crc kubenswrapper[4930]: I0313 09:30:35.605644 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" podStartSLOduration=2.9523782929999998 podStartE2EDuration="6.605627501s" podCreationTimestamp="2026-03-13 09:30:29 +0000 UTC" firstStartedPulling="2026-03-13 09:30:30.907089232 +0000 UTC m=+1071.657003909" lastFinishedPulling="2026-03-13 09:30:34.56033845 +0000 UTC m=+1075.310253117" observedRunningTime="2026-03-13 09:30:35.604313079 +0000 UTC m=+1076.354227756" watchObservedRunningTime="2026-03-13 09:30:35.605627501 +0000 UTC m=+1076.355542178" Mar 13 09:30:40 crc kubenswrapper[4930]: I0313 09:30:40.316904 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" Mar 13 09:30:41 crc kubenswrapper[4930]: I0313 09:30:41.601824 4930 scope.go:117] "RemoveContainer" containerID="65aface3ab07f570391b70eada9c025d734856e880f328929b9dd4abe2dff180" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.004732 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.006688 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.009474 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-xgt7j" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.016896 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.017919 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.019309 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-ht84s" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.030906 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.048706 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.054344 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.055560 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.060333 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-xdcd4" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.095674 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.096739 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.105395 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.106661 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.111214 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-f5xvl" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.111671 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-9frk6" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.124777 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.131916 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jssb2\" (UniqueName: \"kubernetes.io/projected/2e04d1ac-077a-4dc5-91a4-c9b22c352ac2-kube-api-access-jssb2\") pod \"barbican-operator-controller-manager-677bd678f7-kckp5\" (UID: \"2e04d1ac-077a-4dc5-91a4-c9b22c352ac2\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.132417 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnlxj\" (UniqueName: \"kubernetes.io/projected/4051662d-ae7a-4f38-9aa5-5eecab9693f3-kube-api-access-fnlxj\") pod \"designate-operator-controller-manager-66d56f6ff4-nsxgv\" (UID: \"4051662d-ae7a-4f38-9aa5-5eecab9693f3\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.132469 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtrlv\" (UniqueName: \"kubernetes.io/projected/fd7bda25-ce82-4cc8-bc80-ecd97b9622cc-kube-api-access-gtrlv\") pod \"cinder-operator-controller-manager-984cd4dcf-mcqf7\" (UID: \"fd7bda25-ce82-4cc8-bc80-ecd97b9622cc\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.132546 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmktv\" (UniqueName: \"kubernetes.io/projected/4a2edceb-a605-4edd-8905-e35177c281c7-kube-api-access-vmktv\") pod \"heat-operator-controller-manager-77b6666d85-npvw9\" (UID: \"4a2edceb-a605-4edd-8905-e35177c281c7\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.136585 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.174511 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.236471 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8k7rw\" (UniqueName: \"kubernetes.io/projected/ee0102ec-b840-47fb-bae7-4b6e04c231fa-kube-api-access-8k7rw\") pod \"glance-operator-controller-manager-5964f64c48-6vcbx\" (UID: \"ee0102ec-b840-47fb-bae7-4b6e04c231fa\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.236560 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmktv\" (UniqueName: \"kubernetes.io/projected/4a2edceb-a605-4edd-8905-e35177c281c7-kube-api-access-vmktv\") pod \"heat-operator-controller-manager-77b6666d85-npvw9\" (UID: \"4a2edceb-a605-4edd-8905-e35177c281c7\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.236745 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jssb2\" (UniqueName: \"kubernetes.io/projected/2e04d1ac-077a-4dc5-91a4-c9b22c352ac2-kube-api-access-jssb2\") pod \"barbican-operator-controller-manager-677bd678f7-kckp5\" (UID: \"2e04d1ac-077a-4dc5-91a4-c9b22c352ac2\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.236796 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnlxj\" (UniqueName: \"kubernetes.io/projected/4051662d-ae7a-4f38-9aa5-5eecab9693f3-kube-api-access-fnlxj\") pod \"designate-operator-controller-manager-66d56f6ff4-nsxgv\" (UID: \"4051662d-ae7a-4f38-9aa5-5eecab9693f3\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.236821 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtrlv\" (UniqueName: \"kubernetes.io/projected/fd7bda25-ce82-4cc8-bc80-ecd97b9622cc-kube-api-access-gtrlv\") pod \"cinder-operator-controller-manager-984cd4dcf-mcqf7\" (UID: \"fd7bda25-ce82-4cc8-bc80-ecd97b9622cc\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.241510 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.242614 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.253982 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-5h27s" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.277986 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.279122 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.292458 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jssb2\" (UniqueName: \"kubernetes.io/projected/2e04d1ac-077a-4dc5-91a4-c9b22c352ac2-kube-api-access-jssb2\") pod \"barbican-operator-controller-manager-677bd678f7-kckp5\" (UID: \"2e04d1ac-077a-4dc5-91a4-c9b22c352ac2\") " pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.301490 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.301502 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmktv\" (UniqueName: \"kubernetes.io/projected/4a2edceb-a605-4edd-8905-e35177c281c7-kube-api-access-vmktv\") pod \"heat-operator-controller-manager-77b6666d85-npvw9\" (UID: \"4a2edceb-a605-4edd-8905-e35177c281c7\") " pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.301738 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-rw2pf" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.302206 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnlxj\" (UniqueName: \"kubernetes.io/projected/4051662d-ae7a-4f38-9aa5-5eecab9693f3-kube-api-access-fnlxj\") pod \"designate-operator-controller-manager-66d56f6ff4-nsxgv\" (UID: \"4051662d-ae7a-4f38-9aa5-5eecab9693f3\") " pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.302255 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.304127 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtrlv\" (UniqueName: \"kubernetes.io/projected/fd7bda25-ce82-4cc8-bc80-ecd97b9622cc-kube-api-access-gtrlv\") pod \"cinder-operator-controller-manager-984cd4dcf-mcqf7\" (UID: \"fd7bda25-ce82-4cc8-bc80-ecd97b9622cc\") " pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.319462 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.334091 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.343413 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.345266 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8k7rw\" (UniqueName: \"kubernetes.io/projected/ee0102ec-b840-47fb-bae7-4b6e04c231fa-kube-api-access-8k7rw\") pod \"glance-operator-controller-manager-5964f64c48-6vcbx\" (UID: \"ee0102ec-b840-47fb-bae7-4b6e04c231fa\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.345451 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvfkj\" (UniqueName: \"kubernetes.io/projected/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-kube-api-access-hvfkj\") pod \"infra-operator-controller-manager-5995f4446f-mqvs5\" (UID: \"28acf0f4-52a0-40f4-9a05-abbdcc79fb33\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.345538 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqhs8\" (UniqueName: \"kubernetes.io/projected/12a9ad60-3dff-443d-9f09-b57c4a373aa2-kube-api-access-dqhs8\") pod \"horizon-operator-controller-manager-6d9d6b584d-xlwng\" (UID: \"12a9ad60-3dff-443d-9f09-b57c4a373aa2\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.345623 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert\") pod \"infra-operator-controller-manager-5995f4446f-mqvs5\" (UID: \"28acf0f4-52a0-40f4-9a05-abbdcc79fb33\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.346485 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.347839 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.358826 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-pdncv" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.360188 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.377336 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.378942 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8k7rw\" (UniqueName: \"kubernetes.io/projected/ee0102ec-b840-47fb-bae7-4b6e04c231fa-kube-api-access-8k7rw\") pod \"glance-operator-controller-manager-5964f64c48-6vcbx\" (UID: \"ee0102ec-b840-47fb-bae7-4b6e04c231fa\") " pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.381870 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.383406 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-sml6g" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.384469 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.394993 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.396125 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.397704 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-d6f55" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.399795 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.401674 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.412795 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.414149 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.425784 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-2hbd7" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.431818 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.436867 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.437923 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.446233 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-4z5zz" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.446687 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.447625 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7kg9\" (UniqueName: \"kubernetes.io/projected/0d606b70-8e59-4303-963c-54c0be0a3800-kube-api-access-v7kg9\") pod \"mariadb-operator-controller-manager-658d4cdd5-tk79c\" (UID: \"0d606b70-8e59-4303-963c-54c0be0a3800\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.447702 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5k94\" (UniqueName: \"kubernetes.io/projected/d2ea88fb-18e6-4297-bb18-48d0e1d566be-kube-api-access-j5k94\") pod \"ironic-operator-controller-manager-6bbb499bbc-7xbzn\" (UID: \"d2ea88fb-18e6-4297-bb18-48d0e1d566be\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.447734 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq942\" (UniqueName: \"kubernetes.io/projected/5ede266e-f821-4c72-a408-16b4b71a483e-kube-api-access-jq942\") pod \"manila-operator-controller-manager-68f45f9d9f-vsnn4\" (UID: \"5ede266e-f821-4c72-a408-16b4b71a483e\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.447762 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvfkj\" (UniqueName: \"kubernetes.io/projected/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-kube-api-access-hvfkj\") pod \"infra-operator-controller-manager-5995f4446f-mqvs5\" (UID: \"28acf0f4-52a0-40f4-9a05-abbdcc79fb33\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.447783 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9ps2p\" (UniqueName: \"kubernetes.io/projected/47a82845-cf9f-4fc8-a35f-4548a9c7127e-kube-api-access-9ps2p\") pod \"keystone-operator-controller-manager-684f77d66d-rmm4c\" (UID: \"47a82845-cf9f-4fc8-a35f-4548a9c7127e\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.447803 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqhs8\" (UniqueName: \"kubernetes.io/projected/12a9ad60-3dff-443d-9f09-b57c4a373aa2-kube-api-access-dqhs8\") pod \"horizon-operator-controller-manager-6d9d6b584d-xlwng\" (UID: \"12a9ad60-3dff-443d-9f09-b57c4a373aa2\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.447831 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert\") pod \"infra-operator-controller-manager-5995f4446f-mqvs5\" (UID: \"28acf0f4-52a0-40f4-9a05-abbdcc79fb33\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:00 crc kubenswrapper[4930]: E0313 09:31:00.447996 4930 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 09:31:00 crc kubenswrapper[4930]: E0313 09:31:00.448057 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert podName:28acf0f4-52a0-40f4-9a05-abbdcc79fb33 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:00.948039757 +0000 UTC m=+1101.697954424 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert") pod "infra-operator-controller-manager-5995f4446f-mqvs5" (UID: "28acf0f4-52a0-40f4-9a05-abbdcc79fb33") : secret "infra-operator-webhook-server-cert" not found Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.455765 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.456394 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.461567 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.462982 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.472490 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.473633 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.476796 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-p9gql" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.477683 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-hlkzm" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.484269 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.488227 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqhs8\" (UniqueName: \"kubernetes.io/projected/12a9ad60-3dff-443d-9f09-b57c4a373aa2-kube-api-access-dqhs8\") pod \"horizon-operator-controller-manager-6d9d6b584d-xlwng\" (UID: \"12a9ad60-3dff-443d-9f09-b57c4a373aa2\") " pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.491456 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvfkj\" (UniqueName: \"kubernetes.io/projected/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-kube-api-access-hvfkj\") pod \"infra-operator-controller-manager-5995f4446f-mqvs5\" (UID: \"28acf0f4-52a0-40f4-9a05-abbdcc79fb33\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.495809 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.513104 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.523676 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.525017 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.526969 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-7s24r" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.527189 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.547905 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.549100 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.551076 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5k94\" (UniqueName: \"kubernetes.io/projected/d2ea88fb-18e6-4297-bb18-48d0e1d566be-kube-api-access-j5k94\") pod \"ironic-operator-controller-manager-6bbb499bbc-7xbzn\" (UID: \"d2ea88fb-18e6-4297-bb18-48d0e1d566be\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.551381 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq942\" (UniqueName: \"kubernetes.io/projected/5ede266e-f821-4c72-a408-16b4b71a483e-kube-api-access-jq942\") pod \"manila-operator-controller-manager-68f45f9d9f-vsnn4\" (UID: \"5ede266e-f821-4c72-a408-16b4b71a483e\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.551413 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pfggs\" (UniqueName: \"kubernetes.io/projected/561d468b-4138-4196-8652-38b9eb4519ea-kube-api-access-pfggs\") pod \"octavia-operator-controller-manager-5f4f55cb5c-qx24r\" (UID: \"561d468b-4138-4196-8652-38b9eb4519ea\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.551457 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9ps2p\" (UniqueName: \"kubernetes.io/projected/47a82845-cf9f-4fc8-a35f-4548a9c7127e-kube-api-access-9ps2p\") pod \"keystone-operator-controller-manager-684f77d66d-rmm4c\" (UID: \"47a82845-cf9f-4fc8-a35f-4548a9c7127e\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.551487 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfv9k\" (UniqueName: \"kubernetes.io/projected/d12dd258-eed7-41ff-b2e9-f124e189e9aa-kube-api-access-lfv9k\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw\" (UID: \"d12dd258-eed7-41ff-b2e9-f124e189e9aa\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.551554 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z65t6\" (UniqueName: \"kubernetes.io/projected/585f7875-1138-4fcb-b89a-0a5f208f04b7-kube-api-access-z65t6\") pod \"nova-operator-controller-manager-569cc54c5-wwdw2\" (UID: \"585f7875-1138-4fcb-b89a-0a5f208f04b7\") " pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.551579 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw\" (UID: \"d12dd258-eed7-41ff-b2e9-f124e189e9aa\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.551600 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7kg9\" (UniqueName: \"kubernetes.io/projected/0d606b70-8e59-4303-963c-54c0be0a3800-kube-api-access-v7kg9\") pod \"mariadb-operator-controller-manager-658d4cdd5-tk79c\" (UID: \"0d606b70-8e59-4303-963c-54c0be0a3800\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.551644 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvbpx\" (UniqueName: \"kubernetes.io/projected/a28dc996-b000-4680-b9f1-d770e91f0cba-kube-api-access-jvbpx\") pod \"neutron-operator-controller-manager-776c5696bf-545tf\" (UID: \"a28dc996-b000-4680-b9f1-d770e91f0cba\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.563493 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.564870 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.569910 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-hpf9n" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.570254 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-w7hwt" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.579073 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7kg9\" (UniqueName: \"kubernetes.io/projected/0d606b70-8e59-4303-963c-54c0be0a3800-kube-api-access-v7kg9\") pod \"mariadb-operator-controller-manager-658d4cdd5-tk79c\" (UID: \"0d606b70-8e59-4303-963c-54c0be0a3800\") " pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.580355 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5k94\" (UniqueName: \"kubernetes.io/projected/d2ea88fb-18e6-4297-bb18-48d0e1d566be-kube-api-access-j5k94\") pod \"ironic-operator-controller-manager-6bbb499bbc-7xbzn\" (UID: \"d2ea88fb-18e6-4297-bb18-48d0e1d566be\") " pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.580638 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9ps2p\" (UniqueName: \"kubernetes.io/projected/47a82845-cf9f-4fc8-a35f-4548a9c7127e-kube-api-access-9ps2p\") pod \"keystone-operator-controller-manager-684f77d66d-rmm4c\" (UID: \"47a82845-cf9f-4fc8-a35f-4548a9c7127e\") " pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.583510 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.583993 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq942\" (UniqueName: \"kubernetes.io/projected/5ede266e-f821-4c72-a408-16b4b71a483e-kube-api-access-jq942\") pod \"manila-operator-controller-manager-68f45f9d9f-vsnn4\" (UID: \"5ede266e-f821-4c72-a408-16b4b71a483e\") " pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.584424 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.599614 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.627791 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.647485 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.654359 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw\" (UID: \"d12dd258-eed7-41ff-b2e9-f124e189e9aa\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:00 crc kubenswrapper[4930]: E0313 09:31:00.654527 4930 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.654538 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqcz7\" (UniqueName: \"kubernetes.io/projected/046b75aa-f3a0-467d-8624-6a82aa748b32-kube-api-access-gqcz7\") pod \"placement-operator-controller-manager-574d45c66c-zrr4m\" (UID: \"046b75aa-f3a0-467d-8624-6a82aa748b32\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" Mar 13 09:31:00 crc kubenswrapper[4930]: E0313 09:31:00.654581 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert podName:d12dd258-eed7-41ff-b2e9-f124e189e9aa nodeName:}" failed. No retries permitted until 2026-03-13 09:31:01.154564305 +0000 UTC m=+1101.904478982 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" (UID: "d12dd258-eed7-41ff-b2e9-f124e189e9aa") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.654602 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvbpx\" (UniqueName: \"kubernetes.io/projected/a28dc996-b000-4680-b9f1-d770e91f0cba-kube-api-access-jvbpx\") pod \"neutron-operator-controller-manager-776c5696bf-545tf\" (UID: \"a28dc996-b000-4680-b9f1-d770e91f0cba\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.654802 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pv4gt\" (UniqueName: \"kubernetes.io/projected/6c666666-f9ed-4183-9a92-e0334239aa3e-kube-api-access-pv4gt\") pod \"ovn-operator-controller-manager-bbc5b68f9-f45t2\" (UID: \"6c666666-f9ed-4183-9a92-e0334239aa3e\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.654856 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pfggs\" (UniqueName: \"kubernetes.io/projected/561d468b-4138-4196-8652-38b9eb4519ea-kube-api-access-pfggs\") pod \"octavia-operator-controller-manager-5f4f55cb5c-qx24r\" (UID: \"561d468b-4138-4196-8652-38b9eb4519ea\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.654897 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfv9k\" (UniqueName: \"kubernetes.io/projected/d12dd258-eed7-41ff-b2e9-f124e189e9aa-kube-api-access-lfv9k\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw\" (UID: \"d12dd258-eed7-41ff-b2e9-f124e189e9aa\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.654954 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z65t6\" (UniqueName: \"kubernetes.io/projected/585f7875-1138-4fcb-b89a-0a5f208f04b7-kube-api-access-z65t6\") pod \"nova-operator-controller-manager-569cc54c5-wwdw2\" (UID: \"585f7875-1138-4fcb-b89a-0a5f208f04b7\") " pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.683896 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.690031 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.696822 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvbpx\" (UniqueName: \"kubernetes.io/projected/a28dc996-b000-4680-b9f1-d770e91f0cba-kube-api-access-jvbpx\") pod \"neutron-operator-controller-manager-776c5696bf-545tf\" (UID: \"a28dc996-b000-4680-b9f1-d770e91f0cba\") " pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.701821 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z65t6\" (UniqueName: \"kubernetes.io/projected/585f7875-1138-4fcb-b89a-0a5f208f04b7-kube-api-access-z65t6\") pod \"nova-operator-controller-manager-569cc54c5-wwdw2\" (UID: \"585f7875-1138-4fcb-b89a-0a5f208f04b7\") " pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.703201 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.706087 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.707639 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.714553 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.717568 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pfggs\" (UniqueName: \"kubernetes.io/projected/561d468b-4138-4196-8652-38b9eb4519ea-kube-api-access-pfggs\") pod \"octavia-operator-controller-manager-5f4f55cb5c-qx24r\" (UID: \"561d468b-4138-4196-8652-38b9eb4519ea\") " pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.717747 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfv9k\" (UniqueName: \"kubernetes.io/projected/d12dd258-eed7-41ff-b2e9-f124e189e9aa-kube-api-access-lfv9k\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw\" (UID: \"d12dd258-eed7-41ff-b2e9-f124e189e9aa\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.718066 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-mn5zl" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.732170 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.760830 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pv4gt\" (UniqueName: \"kubernetes.io/projected/6c666666-f9ed-4183-9a92-e0334239aa3e-kube-api-access-pv4gt\") pod \"ovn-operator-controller-manager-bbc5b68f9-f45t2\" (UID: \"6c666666-f9ed-4183-9a92-e0334239aa3e\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.761068 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqcz7\" (UniqueName: \"kubernetes.io/projected/046b75aa-f3a0-467d-8624-6a82aa748b32-kube-api-access-gqcz7\") pod \"placement-operator-controller-manager-574d45c66c-zrr4m\" (UID: \"046b75aa-f3a0-467d-8624-6a82aa748b32\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.767463 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.788093 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.804259 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.807482 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-cnfkw" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.810216 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqcz7\" (UniqueName: \"kubernetes.io/projected/046b75aa-f3a0-467d-8624-6a82aa748b32-kube-api-access-gqcz7\") pod \"placement-operator-controller-manager-574d45c66c-zrr4m\" (UID: \"046b75aa-f3a0-467d-8624-6a82aa748b32\") " pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.813205 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pv4gt\" (UniqueName: \"kubernetes.io/projected/6c666666-f9ed-4183-9a92-e0334239aa3e-kube-api-access-pv4gt\") pod \"ovn-operator-controller-manager-bbc5b68f9-f45t2\" (UID: \"6c666666-f9ed-4183-9a92-e0334239aa3e\") " pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.816604 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.862498 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.869807 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56c7q\" (UniqueName: \"kubernetes.io/projected/1a25b1bf-af64-4f8c-920c-c450ebc66b41-kube-api-access-56c7q\") pod \"telemetry-operator-controller-manager-5ffc4556d7-9h5cs\" (UID: \"1a25b1bf-af64-4f8c-920c-c450ebc66b41\") " pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.869865 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twplw\" (UniqueName: \"kubernetes.io/projected/7601eb51-d514-4a96-848c-af0eae32f54f-kube-api-access-twplw\") pod \"swift-operator-controller-manager-677c674df7-ntwzc\" (UID: \"7601eb51-d514-4a96-848c-af0eae32f54f\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.874003 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.947539 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.949220 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.951908 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-pqzdw" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.971910 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert\") pod \"infra-operator-controller-manager-5995f4446f-mqvs5\" (UID: \"28acf0f4-52a0-40f4-9a05-abbdcc79fb33\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.971996 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56c7q\" (UniqueName: \"kubernetes.io/projected/1a25b1bf-af64-4f8c-920c-c450ebc66b41-kube-api-access-56c7q\") pod \"telemetry-operator-controller-manager-5ffc4556d7-9h5cs\" (UID: \"1a25b1bf-af64-4f8c-920c-c450ebc66b41\") " pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.972028 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twplw\" (UniqueName: \"kubernetes.io/projected/7601eb51-d514-4a96-848c-af0eae32f54f-kube-api-access-twplw\") pod \"swift-operator-controller-manager-677c674df7-ntwzc\" (UID: \"7601eb51-d514-4a96-848c-af0eae32f54f\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" Mar 13 09:31:00 crc kubenswrapper[4930]: E0313 09:31:00.972594 4930 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 09:31:00 crc kubenswrapper[4930]: E0313 09:31:00.972635 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert podName:28acf0f4-52a0-40f4-9a05-abbdcc79fb33 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:01.972620471 +0000 UTC m=+1102.722535148 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert") pod "infra-operator-controller-manager-5995f4446f-mqvs5" (UID: "28acf0f4-52a0-40f4-9a05-abbdcc79fb33") : secret "infra-operator-webhook-server-cert" not found Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.978605 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f"] Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.993712 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twplw\" (UniqueName: \"kubernetes.io/projected/7601eb51-d514-4a96-848c-af0eae32f54f-kube-api-access-twplw\") pod \"swift-operator-controller-manager-677c674df7-ntwzc\" (UID: \"7601eb51-d514-4a96-848c-af0eae32f54f\") " pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" Mar 13 09:31:00 crc kubenswrapper[4930]: I0313 09:31:00.995603 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56c7q\" (UniqueName: \"kubernetes.io/projected/1a25b1bf-af64-4f8c-920c-c450ebc66b41-kube-api-access-56c7q\") pod \"telemetry-operator-controller-manager-5ffc4556d7-9h5cs\" (UID: \"1a25b1bf-af64-4f8c-920c-c450ebc66b41\") " pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.032204 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm"] Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.033205 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.035226 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-zb4kb" Mar 13 09:31:01 crc kubenswrapper[4930]: W0313 09:31:01.041814 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2e04d1ac_077a_4dc5_91a4_c9b22c352ac2.slice/crio-547394ae47eb021b567b45f28c4876fedb0c19cc3f5c6a52ebd856db1f141b31 WatchSource:0}: Error finding container 547394ae47eb021b567b45f28c4876fedb0c19cc3f5c6a52ebd856db1f141b31: Status 404 returned error can't find the container with id 547394ae47eb021b567b45f28c4876fedb0c19cc3f5c6a52ebd856db1f141b31 Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.059392 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm"] Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.074424 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vb5mk\" (UniqueName: \"kubernetes.io/projected/c92ab5a7-1ad7-4584-9954-0047b879bc47-kube-api-access-vb5mk\") pod \"test-operator-controller-manager-5c5cb9c4d7-8xc7f\" (UID: \"c92ab5a7-1ad7-4584-9954-0047b879bc47\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.085494 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8"] Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.086822 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.091001 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-ll7rf" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.091158 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.091260 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.095780 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8"] Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.121645 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc"] Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.123186 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.124997 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-9rl4f" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.138744 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc"] Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.166344 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5"] Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.177784 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.177871 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nlbgn\" (UniqueName: \"kubernetes.io/projected/eaa30848-0297-4cdc-8c48-379fe75ad943-kube-api-access-nlbgn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-wdbjc\" (UID: \"eaa30848-0297-4cdc-8c48-379fe75ad943\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.177953 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrng9\" (UniqueName: \"kubernetes.io/projected/984e8970-7f0d-4ce8-87c2-56c787857225-kube-api-access-rrng9\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.177998 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw\" (UID: \"d12dd258-eed7-41ff-b2e9-f124e189e9aa\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.178089 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vb5mk\" (UniqueName: \"kubernetes.io/projected/c92ab5a7-1ad7-4584-9954-0047b879bc47-kube-api-access-vb5mk\") pod \"test-operator-controller-manager-5c5cb9c4d7-8xc7f\" (UID: \"c92ab5a7-1ad7-4584-9954-0047b879bc47\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.178146 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hmfs\" (UniqueName: \"kubernetes.io/projected/43277746-d9dc-4452-bcaf-8c4668ebf473-kube-api-access-2hmfs\") pod \"watcher-operator-controller-manager-6dd88c6f67-2jptm\" (UID: \"43277746-d9dc-4452-bcaf-8c4668ebf473\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.178177 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:01 crc kubenswrapper[4930]: E0313 09:31:01.178358 4930 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 09:31:01 crc kubenswrapper[4930]: E0313 09:31:01.178419 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert podName:d12dd258-eed7-41ff-b2e9-f124e189e9aa nodeName:}" failed. No retries permitted until 2026-03-13 09:31:02.17839825 +0000 UTC m=+1102.928312927 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" (UID: "d12dd258-eed7-41ff-b2e9-f124e189e9aa") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.191494 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.199485 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vb5mk\" (UniqueName: \"kubernetes.io/projected/c92ab5a7-1ad7-4584-9954-0047b879bc47-kube-api-access-vb5mk\") pod \"test-operator-controller-manager-5c5cb9c4d7-8xc7f\" (UID: \"c92ab5a7-1ad7-4584-9954-0047b879bc47\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.225076 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.277980 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.278808 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.278871 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nlbgn\" (UniqueName: \"kubernetes.io/projected/eaa30848-0297-4cdc-8c48-379fe75ad943-kube-api-access-nlbgn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-wdbjc\" (UID: \"eaa30848-0297-4cdc-8c48-379fe75ad943\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.278940 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrng9\" (UniqueName: \"kubernetes.io/projected/984e8970-7f0d-4ce8-87c2-56c787857225-kube-api-access-rrng9\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.279009 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hmfs\" (UniqueName: \"kubernetes.io/projected/43277746-d9dc-4452-bcaf-8c4668ebf473-kube-api-access-2hmfs\") pod \"watcher-operator-controller-manager-6dd88c6f67-2jptm\" (UID: \"43277746-d9dc-4452-bcaf-8c4668ebf473\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.279031 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:01 crc kubenswrapper[4930]: E0313 09:31:01.279210 4930 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 09:31:01 crc kubenswrapper[4930]: E0313 09:31:01.279259 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs podName:984e8970-7f0d-4ce8-87c2-56c787857225 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:01.779244013 +0000 UTC m=+1102.529158690 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs") pod "openstack-operator-controller-manager-6d5df8c674-hkqw8" (UID: "984e8970-7f0d-4ce8-87c2-56c787857225") : secret "metrics-server-cert" not found Mar 13 09:31:01 crc kubenswrapper[4930]: E0313 09:31:01.279821 4930 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 09:31:01 crc kubenswrapper[4930]: E0313 09:31:01.279881 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs podName:984e8970-7f0d-4ce8-87c2-56c787857225 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:01.779859919 +0000 UTC m=+1102.529774686 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs") pod "openstack-operator-controller-manager-6d5df8c674-hkqw8" (UID: "984e8970-7f0d-4ce8-87c2-56c787857225") : secret "webhook-server-cert" not found Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.304378 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrng9\" (UniqueName: \"kubernetes.io/projected/984e8970-7f0d-4ce8-87c2-56c787857225-kube-api-access-rrng9\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.306081 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hmfs\" (UniqueName: \"kubernetes.io/projected/43277746-d9dc-4452-bcaf-8c4668ebf473-kube-api-access-2hmfs\") pod \"watcher-operator-controller-manager-6dd88c6f67-2jptm\" (UID: \"43277746-d9dc-4452-bcaf-8c4668ebf473\") " pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.319163 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nlbgn\" (UniqueName: \"kubernetes.io/projected/eaa30848-0297-4cdc-8c48-379fe75ad943-kube-api-access-nlbgn\") pod \"rabbitmq-cluster-operator-manager-668c99d594-wdbjc\" (UID: \"eaa30848-0297-4cdc-8c48-379fe75ad943\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.352692 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.429508 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv"] Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.437062 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7"] Mar 13 09:31:01 crc kubenswrapper[4930]: W0313 09:31:01.450794 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfd7bda25_ce82_4cc8_bc80_ecd97b9622cc.slice/crio-11dfe054449e43425a714920e7b6897f534fa61e3eb215dc1559118afbdc7e8e WatchSource:0}: Error finding container 11dfe054449e43425a714920e7b6897f534fa61e3eb215dc1559118afbdc7e8e: Status 404 returned error can't find the container with id 11dfe054449e43425a714920e7b6897f534fa61e3eb215dc1559118afbdc7e8e Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.470713 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" Mar 13 09:31:01 crc kubenswrapper[4930]: W0313 09:31:01.471185 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4051662d_ae7a_4f38_9aa5_5eecab9693f3.slice/crio-adb136ca916687e36e8e23e257d2d552449dd8ba072195ed128891cdfa263d63 WatchSource:0}: Error finding container adb136ca916687e36e8e23e257d2d552449dd8ba072195ed128891cdfa263d63: Status 404 returned error can't find the container with id adb136ca916687e36e8e23e257d2d552449dd8ba072195ed128891cdfa263d63 Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.719175 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9"] Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.733134 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx"] Mar 13 09:31:01 crc kubenswrapper[4930]: W0313 09:31:01.749805 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd2ea88fb_18e6_4297_bb18_48d0e1d566be.slice/crio-dc1034a680f48e1f71a7ac3fc04163a54adb17a3a907deefbfd9c33d5fce5df9 WatchSource:0}: Error finding container dc1034a680f48e1f71a7ac3fc04163a54adb17a3a907deefbfd9c33d5fce5df9: Status 404 returned error can't find the container with id dc1034a680f48e1f71a7ac3fc04163a54adb17a3a907deefbfd9c33d5fce5df9 Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.758420 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn"] Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.791730 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.791822 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:01 crc kubenswrapper[4930]: E0313 09:31:01.791942 4930 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 09:31:01 crc kubenswrapper[4930]: E0313 09:31:01.792016 4930 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 09:31:01 crc kubenswrapper[4930]: E0313 09:31:01.792021 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs podName:984e8970-7f0d-4ce8-87c2-56c787857225 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:02.792001182 +0000 UTC m=+1103.541915849 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs") pod "openstack-operator-controller-manager-6d5df8c674-hkqw8" (UID: "984e8970-7f0d-4ce8-87c2-56c787857225") : secret "metrics-server-cert" not found Mar 13 09:31:01 crc kubenswrapper[4930]: E0313 09:31:01.792091 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs podName:984e8970-7f0d-4ce8-87c2-56c787857225 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:02.792072793 +0000 UTC m=+1103.541987580 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs") pod "openstack-operator-controller-manager-6d5df8c674-hkqw8" (UID: "984e8970-7f0d-4ce8-87c2-56c787857225") : secret "webhook-server-cert" not found Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.835534 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" event={"ID":"ee0102ec-b840-47fb-bae7-4b6e04c231fa","Type":"ContainerStarted","Data":"892908c286d551de29fd6e4777ebfc9f989b061e1f70bb77f40a0660fbef4874"} Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.836984 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" event={"ID":"2e04d1ac-077a-4dc5-91a4-c9b22c352ac2","Type":"ContainerStarted","Data":"547394ae47eb021b567b45f28c4876fedb0c19cc3f5c6a52ebd856db1f141b31"} Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.841151 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" event={"ID":"4051662d-ae7a-4f38-9aa5-5eecab9693f3","Type":"ContainerStarted","Data":"adb136ca916687e36e8e23e257d2d552449dd8ba072195ed128891cdfa263d63"} Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.842348 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" event={"ID":"fd7bda25-ce82-4cc8-bc80-ecd97b9622cc","Type":"ContainerStarted","Data":"11dfe054449e43425a714920e7b6897f534fa61e3eb215dc1559118afbdc7e8e"} Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.843644 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" event={"ID":"4a2edceb-a605-4edd-8905-e35177c281c7","Type":"ContainerStarted","Data":"95c2221670cc6ed180db13ef4604269d9dde8bc6af7e0dd48ae1dabfda761f0d"} Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.844922 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" event={"ID":"d2ea88fb-18e6-4297-bb18-48d0e1d566be","Type":"ContainerStarted","Data":"dc1034a680f48e1f71a7ac3fc04163a54adb17a3a907deefbfd9c33d5fce5df9"} Mar 13 09:31:01 crc kubenswrapper[4930]: I0313 09:31:01.993746 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert\") pod \"infra-operator-controller-manager-5995f4446f-mqvs5\" (UID: \"28acf0f4-52a0-40f4-9a05-abbdcc79fb33\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:01 crc kubenswrapper[4930]: E0313 09:31:01.993876 4930 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 09:31:01 crc kubenswrapper[4930]: E0313 09:31:01.993920 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert podName:28acf0f4-52a0-40f4-9a05-abbdcc79fb33 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:03.993907654 +0000 UTC m=+1104.743822331 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert") pod "infra-operator-controller-manager-5995f4446f-mqvs5" (UID: "28acf0f4-52a0-40f4-9a05-abbdcc79fb33") : secret "infra-operator-webhook-server-cert" not found Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.198347 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw\" (UID: \"d12dd258-eed7-41ff-b2e9-f124e189e9aa\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:02 crc kubenswrapper[4930]: E0313 09:31:02.198471 4930 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 09:31:02 crc kubenswrapper[4930]: E0313 09:31:02.198558 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert podName:d12dd258-eed7-41ff-b2e9-f124e189e9aa nodeName:}" failed. No retries permitted until 2026-03-13 09:31:04.198518253 +0000 UTC m=+1104.948432930 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" (UID: "d12dd258-eed7-41ff-b2e9-f124e189e9aa") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.357989 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2"] Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.386514 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c"] Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.406966 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf"] Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.428697 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r"] Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.436515 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng"] Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.546497 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2"] Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.576373 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4"] Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.599560 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c"] Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.827618 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f"] Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.834828 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.834909 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:02 crc kubenswrapper[4930]: E0313 09:31:02.835052 4930 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 09:31:02 crc kubenswrapper[4930]: E0313 09:31:02.835136 4930 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 09:31:02 crc kubenswrapper[4930]: E0313 09:31:02.835184 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs podName:984e8970-7f0d-4ce8-87c2-56c787857225 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:04.83517073 +0000 UTC m=+1105.585085407 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs") pod "openstack-operator-controller-manager-6d5df8c674-hkqw8" (UID: "984e8970-7f0d-4ce8-87c2-56c787857225") : secret "webhook-server-cert" not found Mar 13 09:31:02 crc kubenswrapper[4930]: E0313 09:31:02.835377 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs podName:984e8970-7f0d-4ce8-87c2-56c787857225 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:04.835353885 +0000 UTC m=+1105.585268562 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs") pod "openstack-operator-controller-manager-6d5df8c674-hkqw8" (UID: "984e8970-7f0d-4ce8-87c2-56c787857225") : secret "metrics-server-cert" not found Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.846181 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc"] Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.873089 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" event={"ID":"47a82845-cf9f-4fc8-a35f-4548a9c7127e","Type":"ContainerStarted","Data":"d3a5493743f514b1b6a395f62955ac081806e61b0e7aeb2e1422afdb66e6713d"} Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.874906 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" event={"ID":"561d468b-4138-4196-8652-38b9eb4519ea","Type":"ContainerStarted","Data":"b234df229a7d9cd11ff5abda642a9739e1a4eb4c9dd89495c8133c2665b447a8"} Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.881543 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m"] Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.888531 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" event={"ID":"a28dc996-b000-4680-b9f1-d770e91f0cba","Type":"ContainerStarted","Data":"4ff8f3e0108a7578a4e5749b5f73b411ced54a43eb6c37580bb064edb9740df4"} Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.906261 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" event={"ID":"6c666666-f9ed-4183-9a92-e0334239aa3e","Type":"ContainerStarted","Data":"96b7bda55baf4614d834b1b1423821f6443c3dd9cdf9fcc5559467361ba64bc7"} Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.906607 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm"] Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.911628 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" event={"ID":"0d606b70-8e59-4303-963c-54c0be0a3800","Type":"ContainerStarted","Data":"517ce53dfc918db7092d42123f4294ea451c950c81f4be56ce5ef0172181917c"} Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.915295 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" event={"ID":"c92ab5a7-1ad7-4584-9954-0047b879bc47","Type":"ContainerStarted","Data":"daf538b71d67fff829694bb15f001ea9c75cf13ed53eee638f4ffbef3943fd9e"} Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.921096 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" event={"ID":"585f7875-1138-4fcb-b89a-0a5f208f04b7","Type":"ContainerStarted","Data":"5556f873ab5d73ab424e30cdf22a7c2328299de2bc71ff38124c5b5c9ae043be"} Mar 13 09:31:02 crc kubenswrapper[4930]: E0313 09:31:02.923356 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.103:5001/openstack-k8s-operators/telemetry-operator:1a1a9a719889b8cdda26cbd675f0005643a8f9f2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-56c7q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5ffc4556d7-9h5cs_openstack-operators(1a25b1bf-af64-4f8c-920c-c450ebc66b41): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.924500 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" event={"ID":"5ede266e-f821-4c72-a408-16b4b71a483e","Type":"ContainerStarted","Data":"a1fa6e0b40fe28d82cb06d476eeb6525281df9612bea4cf85c76e4c01fc2a711"} Mar 13 09:31:02 crc kubenswrapper[4930]: E0313 09:31:02.924598 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podUID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.928032 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" event={"ID":"12a9ad60-3dff-443d-9f09-b57c4a373aa2","Type":"ContainerStarted","Data":"8949c4f9290a810a894da843e5c3f3096ccf1f1ff6c4a2a1b2c561ecdb7d25b7"} Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.940832 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs"] Mar 13 09:31:02 crc kubenswrapper[4930]: W0313 09:31:02.943600 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeaa30848_0297_4cdc_8c48_379fe75ad943.slice/crio-719e0e648da1951f4b45a8df25bbd9222fb71f98fb43c794bc083e5fb118e21f WatchSource:0}: Error finding container 719e0e648da1951f4b45a8df25bbd9222fb71f98fb43c794bc083e5fb118e21f: Status 404 returned error can't find the container with id 719e0e648da1951f4b45a8df25bbd9222fb71f98fb43c794bc083e5fb118e21f Mar 13 09:31:02 crc kubenswrapper[4930]: E0313 09:31:02.947793 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nlbgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-wdbjc_openstack-operators(eaa30848-0297-4cdc-8c48-379fe75ad943): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 13 09:31:02 crc kubenswrapper[4930]: E0313 09:31:02.949508 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" podUID="eaa30848-0297-4cdc-8c48-379fe75ad943" Mar 13 09:31:02 crc kubenswrapper[4930]: I0313 09:31:02.954027 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc"] Mar 13 09:31:03 crc kubenswrapper[4930]: I0313 09:31:03.946373 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" event={"ID":"43277746-d9dc-4452-bcaf-8c4668ebf473","Type":"ContainerStarted","Data":"4bd0758da2ba1eee44fe66ba1352280a208fc24bdeaf5eb95eb9dc366b582c86"} Mar 13 09:31:03 crc kubenswrapper[4930]: I0313 09:31:03.951502 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" event={"ID":"1a25b1bf-af64-4f8c-920c-c450ebc66b41","Type":"ContainerStarted","Data":"605639d72786302097d69d7f7efe2aa5b27d1d5884ce1def9ad00dd05a8a8f10"} Mar 13 09:31:03 crc kubenswrapper[4930]: E0313 09:31:03.954150 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.103:5001/openstack-k8s-operators/telemetry-operator:1a1a9a719889b8cdda26cbd675f0005643a8f9f2\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podUID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" Mar 13 09:31:03 crc kubenswrapper[4930]: I0313 09:31:03.958054 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" event={"ID":"046b75aa-f3a0-467d-8624-6a82aa748b32","Type":"ContainerStarted","Data":"b5879d7136cf951a5fdd74d1c16474fbe949f1d8d25e4e8bcf6a6c540df6ac0e"} Mar 13 09:31:03 crc kubenswrapper[4930]: E0313 09:31:03.986499 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" podUID="eaa30848-0297-4cdc-8c48-379fe75ad943" Mar 13 09:31:04 crc kubenswrapper[4930]: I0313 09:31:04.007076 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" event={"ID":"7601eb51-d514-4a96-848c-af0eae32f54f","Type":"ContainerStarted","Data":"35912c9ff0067e4d4b7751a6612dd6281299a724d1245a6b239cc83491c59f9e"} Mar 13 09:31:04 crc kubenswrapper[4930]: I0313 09:31:04.007113 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" event={"ID":"eaa30848-0297-4cdc-8c48-379fe75ad943","Type":"ContainerStarted","Data":"719e0e648da1951f4b45a8df25bbd9222fb71f98fb43c794bc083e5fb118e21f"} Mar 13 09:31:04 crc kubenswrapper[4930]: I0313 09:31:04.083215 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert\") pod \"infra-operator-controller-manager-5995f4446f-mqvs5\" (UID: \"28acf0f4-52a0-40f4-9a05-abbdcc79fb33\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:04 crc kubenswrapper[4930]: E0313 09:31:04.083391 4930 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 09:31:04 crc kubenswrapper[4930]: E0313 09:31:04.083490 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert podName:28acf0f4-52a0-40f4-9a05-abbdcc79fb33 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:08.083462321 +0000 UTC m=+1108.833376988 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert") pod "infra-operator-controller-manager-5995f4446f-mqvs5" (UID: "28acf0f4-52a0-40f4-9a05-abbdcc79fb33") : secret "infra-operator-webhook-server-cert" not found Mar 13 09:31:04 crc kubenswrapper[4930]: I0313 09:31:04.288597 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw\" (UID: \"d12dd258-eed7-41ff-b2e9-f124e189e9aa\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:04 crc kubenswrapper[4930]: E0313 09:31:04.289218 4930 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 09:31:04 crc kubenswrapper[4930]: E0313 09:31:04.289287 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert podName:d12dd258-eed7-41ff-b2e9-f124e189e9aa nodeName:}" failed. No retries permitted until 2026-03-13 09:31:08.2892693 +0000 UTC m=+1109.039183977 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" (UID: "d12dd258-eed7-41ff-b2e9-f124e189e9aa") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 09:31:04 crc kubenswrapper[4930]: I0313 09:31:04.911753 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:04 crc kubenswrapper[4930]: I0313 09:31:04.911830 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:04 crc kubenswrapper[4930]: E0313 09:31:04.911993 4930 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 09:31:04 crc kubenswrapper[4930]: E0313 09:31:04.912288 4930 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 09:31:04 crc kubenswrapper[4930]: E0313 09:31:04.912291 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs podName:984e8970-7f0d-4ce8-87c2-56c787857225 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:08.912055727 +0000 UTC m=+1109.661970404 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs") pod "openstack-operator-controller-manager-6d5df8c674-hkqw8" (UID: "984e8970-7f0d-4ce8-87c2-56c787857225") : secret "metrics-server-cert" not found Mar 13 09:31:04 crc kubenswrapper[4930]: E0313 09:31:04.912350 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs podName:984e8970-7f0d-4ce8-87c2-56c787857225 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:08.912334574 +0000 UTC m=+1109.662249251 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs") pod "openstack-operator-controller-manager-6d5df8c674-hkqw8" (UID: "984e8970-7f0d-4ce8-87c2-56c787857225") : secret "webhook-server-cert" not found Mar 13 09:31:05 crc kubenswrapper[4930]: E0313 09:31:05.001214 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" podUID="eaa30848-0297-4cdc-8c48-379fe75ad943" Mar 13 09:31:05 crc kubenswrapper[4930]: E0313 09:31:05.001294 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.103:5001/openstack-k8s-operators/telemetry-operator:1a1a9a719889b8cdda26cbd675f0005643a8f9f2\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podUID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" Mar 13 09:31:06 crc kubenswrapper[4930]: E0313 09:31:06.003459 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" podUID="eaa30848-0297-4cdc-8c48-379fe75ad943" Mar 13 09:31:08 crc kubenswrapper[4930]: I0313 09:31:08.166992 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert\") pod \"infra-operator-controller-manager-5995f4446f-mqvs5\" (UID: \"28acf0f4-52a0-40f4-9a05-abbdcc79fb33\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:08 crc kubenswrapper[4930]: E0313 09:31:08.167194 4930 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 13 09:31:08 crc kubenswrapper[4930]: E0313 09:31:08.167404 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert podName:28acf0f4-52a0-40f4-9a05-abbdcc79fb33 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:16.167389113 +0000 UTC m=+1116.917303790 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert") pod "infra-operator-controller-manager-5995f4446f-mqvs5" (UID: "28acf0f4-52a0-40f4-9a05-abbdcc79fb33") : secret "infra-operator-webhook-server-cert" not found Mar 13 09:31:08 crc kubenswrapper[4930]: I0313 09:31:08.371092 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw\" (UID: \"d12dd258-eed7-41ff-b2e9-f124e189e9aa\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:08 crc kubenswrapper[4930]: E0313 09:31:08.371243 4930 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 09:31:08 crc kubenswrapper[4930]: E0313 09:31:08.371301 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert podName:d12dd258-eed7-41ff-b2e9-f124e189e9aa nodeName:}" failed. No retries permitted until 2026-03-13 09:31:16.371283953 +0000 UTC m=+1117.121198630 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert") pod "openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" (UID: "d12dd258-eed7-41ff-b2e9-f124e189e9aa") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 13 09:31:08 crc kubenswrapper[4930]: I0313 09:31:08.980668 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:08 crc kubenswrapper[4930]: I0313 09:31:08.980755 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:08 crc kubenswrapper[4930]: E0313 09:31:08.980881 4930 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 09:31:08 crc kubenswrapper[4930]: E0313 09:31:08.980953 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs podName:984e8970-7f0d-4ce8-87c2-56c787857225 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:16.980930689 +0000 UTC m=+1117.730845366 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs") pod "openstack-operator-controller-manager-6d5df8c674-hkqw8" (UID: "984e8970-7f0d-4ce8-87c2-56c787857225") : secret "metrics-server-cert" not found Mar 13 09:31:08 crc kubenswrapper[4930]: E0313 09:31:08.981045 4930 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 09:31:08 crc kubenswrapper[4930]: E0313 09:31:08.981142 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs podName:984e8970-7f0d-4ce8-87c2-56c787857225 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:16.981113394 +0000 UTC m=+1117.731028131 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs") pod "openstack-operator-controller-manager-6d5df8c674-hkqw8" (UID: "984e8970-7f0d-4ce8-87c2-56c787857225") : secret "webhook-server-cert" not found Mar 13 09:31:12 crc kubenswrapper[4930]: I0313 09:31:12.307991 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:31:12 crc kubenswrapper[4930]: I0313 09:31:12.308331 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:31:15 crc kubenswrapper[4930]: E0313 09:31:15.321884 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/manila-operator@sha256:d89f3ca6e909f34d145a880829f5e63f1b6b2d11c520a9c5bea7ed1c30ce38f4" Mar 13 09:31:15 crc kubenswrapper[4930]: E0313 09:31:15.322627 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:d89f3ca6e909f34d145a880829f5e63f1b6b2d11c520a9c5bea7ed1c30ce38f4,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jq942,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-68f45f9d9f-vsnn4_openstack-operators(5ede266e-f821-4c72-a408-16b4b71a483e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:15 crc kubenswrapper[4930]: E0313 09:31:15.324606 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" podUID="5ede266e-f821-4c72-a408-16b4b71a483e" Mar 13 09:31:15 crc kubenswrapper[4930]: E0313 09:31:15.812887 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:b99cd5e08bd85c6aaf717519187ba7bfeea359e1537d43b73a7364b7c38116e2" Mar 13 09:31:15 crc kubenswrapper[4930]: E0313 09:31:15.813153 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:b99cd5e08bd85c6aaf717519187ba7bfeea359e1537d43b73a7364b7c38116e2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-v7kg9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-658d4cdd5-tk79c_openstack-operators(0d606b70-8e59-4303-963c-54c0be0a3800): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:15 crc kubenswrapper[4930]: E0313 09:31:15.815078 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" podUID="0d606b70-8e59-4303-963c-54c0be0a3800" Mar 13 09:31:16 crc kubenswrapper[4930]: E0313 09:31:16.099992 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:d89f3ca6e909f34d145a880829f5e63f1b6b2d11c520a9c5bea7ed1c30ce38f4\\\"\"" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" podUID="5ede266e-f821-4c72-a408-16b4b71a483e" Mar 13 09:31:16 crc kubenswrapper[4930]: E0313 09:31:16.101528 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:b99cd5e08bd85c6aaf717519187ba7bfeea359e1537d43b73a7364b7c38116e2\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" podUID="0d606b70-8e59-4303-963c-54c0be0a3800" Mar 13 09:31:16 crc kubenswrapper[4930]: I0313 09:31:16.219452 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert\") pod \"infra-operator-controller-manager-5995f4446f-mqvs5\" (UID: \"28acf0f4-52a0-40f4-9a05-abbdcc79fb33\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:16 crc kubenswrapper[4930]: I0313 09:31:16.225111 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/28acf0f4-52a0-40f4-9a05-abbdcc79fb33-cert\") pod \"infra-operator-controller-manager-5995f4446f-mqvs5\" (UID: \"28acf0f4-52a0-40f4-9a05-abbdcc79fb33\") " pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:16 crc kubenswrapper[4930]: I0313 09:31:16.422133 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw\" (UID: \"d12dd258-eed7-41ff-b2e9-f124e189e9aa\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:16 crc kubenswrapper[4930]: I0313 09:31:16.425910 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/d12dd258-eed7-41ff-b2e9-f124e189e9aa-cert\") pod \"openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw\" (UID: \"d12dd258-eed7-41ff-b2e9-f124e189e9aa\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:16 crc kubenswrapper[4930]: I0313 09:31:16.434532 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:16 crc kubenswrapper[4930]: E0313 09:31:16.439829 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/swift-operator@sha256:c223309f51714785bd878ad04080f7428567edad793be4f992d492abd77af44c" Mar 13 09:31:16 crc kubenswrapper[4930]: E0313 09:31:16.440062 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:c223309f51714785bd878ad04080f7428567edad793be4f992d492abd77af44c,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-twplw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-677c674df7-ntwzc_openstack-operators(7601eb51-d514-4a96-848c-af0eae32f54f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:16 crc kubenswrapper[4930]: E0313 09:31:16.441533 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" podUID="7601eb51-d514-4a96-848c-af0eae32f54f" Mar 13 09:31:16 crc kubenswrapper[4930]: I0313 09:31:16.445704 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:17 crc kubenswrapper[4930]: I0313 09:31:17.031997 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:17 crc kubenswrapper[4930]: I0313 09:31:17.032120 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:17 crc kubenswrapper[4930]: E0313 09:31:17.032209 4930 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 13 09:31:17 crc kubenswrapper[4930]: E0313 09:31:17.032272 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs podName:984e8970-7f0d-4ce8-87c2-56c787857225 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:33.032254491 +0000 UTC m=+1133.782169178 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs") pod "openstack-operator-controller-manager-6d5df8c674-hkqw8" (UID: "984e8970-7f0d-4ce8-87c2-56c787857225") : secret "metrics-server-cert" not found Mar 13 09:31:17 crc kubenswrapper[4930]: E0313 09:31:17.032275 4930 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 13 09:31:17 crc kubenswrapper[4930]: E0313 09:31:17.032305 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs podName:984e8970-7f0d-4ce8-87c2-56c787857225 nodeName:}" failed. No retries permitted until 2026-03-13 09:31:33.032297032 +0000 UTC m=+1133.782211709 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs") pod "openstack-operator-controller-manager-6d5df8c674-hkqw8" (UID: "984e8970-7f0d-4ce8-87c2-56c787857225") : secret "webhook-server-cert" not found Mar 13 09:31:17 crc kubenswrapper[4930]: E0313 09:31:17.054673 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:2bd37bdd917e3abe72613a734ce5021330242ec8cae9b8da76c57a0765152922" Mar 13 09:31:17 crc kubenswrapper[4930]: E0313 09:31:17.054835 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:2bd37bdd917e3abe72613a734ce5021330242ec8cae9b8da76c57a0765152922,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-z65t6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-569cc54c5-wwdw2_openstack-operators(585f7875-1138-4fcb-b89a-0a5f208f04b7): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:17 crc kubenswrapper[4930]: E0313 09:31:17.056526 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" podUID="585f7875-1138-4fcb-b89a-0a5f208f04b7" Mar 13 09:31:17 crc kubenswrapper[4930]: E0313 09:31:17.111316 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:c223309f51714785bd878ad04080f7428567edad793be4f992d492abd77af44c\\\"\"" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" podUID="7601eb51-d514-4a96-848c-af0eae32f54f" Mar 13 09:31:17 crc kubenswrapper[4930]: E0313 09:31:17.111404 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:2bd37bdd917e3abe72613a734ce5021330242ec8cae9b8da76c57a0765152922\\\"\"" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" podUID="585f7875-1138-4fcb-b89a-0a5f208f04b7" Mar 13 09:31:21 crc kubenswrapper[4930]: E0313 09:31:21.004645 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42" Mar 13 09:31:21 crc kubenswrapper[4930]: E0313 09:31:21.005481 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vb5mk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-8xc7f_openstack-operators(c92ab5a7-1ad7-4584-9954-0047b879bc47): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:21 crc kubenswrapper[4930]: E0313 09:31:21.006628 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" podUID="c92ab5a7-1ad7-4584-9954-0047b879bc47" Mar 13 09:31:21 crc kubenswrapper[4930]: E0313 09:31:21.156033 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" podUID="c92ab5a7-1ad7-4584-9954-0047b879bc47" Mar 13 09:31:21 crc kubenswrapper[4930]: E0313 09:31:21.556485 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:65d0c97340f72a8b23f8e11f4b3efcc6ad37daad9b88e24d4564383a08fa85f7" Mar 13 09:31:21 crc kubenswrapper[4930]: E0313 09:31:21.556642 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:65d0c97340f72a8b23f8e11f4b3efcc6ad37daad9b88e24d4564383a08fa85f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fnlxj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-66d56f6ff4-nsxgv_openstack-operators(4051662d-ae7a-4f38-9aa5-5eecab9693f3): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:21 crc kubenswrapper[4930]: E0313 09:31:21.558184 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" podUID="4051662d-ae7a-4f38-9aa5-5eecab9693f3" Mar 13 09:31:22 crc kubenswrapper[4930]: E0313 09:31:22.163796 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/designate-operator@sha256:65d0c97340f72a8b23f8e11f4b3efcc6ad37daad9b88e24d4564383a08fa85f7\\\"\"" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" podUID="4051662d-ae7a-4f38-9aa5-5eecab9693f3" Mar 13 09:31:23 crc kubenswrapper[4930]: E0313 09:31:23.093744 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:a3bc074ddd9a26d3a8609e5dbdfa85a78449ba1c9b5542bff9949219d6760e60" Mar 13 09:31:23 crc kubenswrapper[4930]: E0313 09:31:23.094230 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:a3bc074ddd9a26d3a8609e5dbdfa85a78449ba1c9b5542bff9949219d6760e60,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8k7rw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5964f64c48-6vcbx_openstack-operators(ee0102ec-b840-47fb-bae7-4b6e04c231fa): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:23 crc kubenswrapper[4930]: E0313 09:31:23.095666 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" podUID="ee0102ec-b840-47fb-bae7-4b6e04c231fa" Mar 13 09:31:23 crc kubenswrapper[4930]: E0313 09:31:23.179277 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/glance-operator@sha256:a3bc074ddd9a26d3a8609e5dbdfa85a78449ba1c9b5542bff9949219d6760e60\\\"\"" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" podUID="ee0102ec-b840-47fb-bae7-4b6e04c231fa" Mar 13 09:31:23 crc kubenswrapper[4930]: E0313 09:31:23.590399 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978" Mar 13 09:31:23 crc kubenswrapper[4930]: E0313 09:31:23.590632 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gqcz7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-574d45c66c-zrr4m_openstack-operators(046b75aa-f3a0-467d-8624-6a82aa748b32): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:23 crc kubenswrapper[4930]: E0313 09:31:23.591845 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" podUID="046b75aa-f3a0-467d-8624-6a82aa748b32" Mar 13 09:31:24 crc kubenswrapper[4930]: E0313 09:31:24.182941 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:e7e865363955c670e41b6c042c4f87abceff78f5495ba5c5c82988baad45c978\\\"\"" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" podUID="046b75aa-f3a0-467d-8624-6a82aa748b32" Mar 13 09:31:26 crc kubenswrapper[4930]: E0313 09:31:26.717972 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca" Mar 13 09:31:26 crc kubenswrapper[4930]: E0313 09:31:26.719178 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9ps2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-684f77d66d-rmm4c_openstack-operators(47a82845-cf9f-4fc8-a35f-4548a9c7127e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:26 crc kubenswrapper[4930]: E0313 09:31:26.720372 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" podUID="47a82845-cf9f-4fc8-a35f-4548a9c7127e" Mar 13 09:31:27 crc kubenswrapper[4930]: E0313 09:31:27.201164 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:40b84319f2f12a1c7ee478fd86a8b1aa5ac2ea8e24f5ce0f1ca78ad879dea8ca\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" podUID="47a82845-cf9f-4fc8-a35f-4548a9c7127e" Mar 13 09:31:27 crc kubenswrapper[4930]: E0313 09:31:27.333029 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/cinder-operator@sha256:7c0da25380c91ffd1940d75eaa71b6842a6a4cf4056e62d6b0d237897b74e4d9" Mar 13 09:31:27 crc kubenswrapper[4930]: E0313 09:31:27.333291 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/cinder-operator@sha256:7c0da25380c91ffd1940d75eaa71b6842a6a4cf4056e62d6b0d237897b74e4d9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gtrlv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-operator-controller-manager-984cd4dcf-mcqf7_openstack-operators(fd7bda25-ce82-4cc8-bc80-ecd97b9622cc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:27 crc kubenswrapper[4930]: E0313 09:31:27.335154 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" podUID="fd7bda25-ce82-4cc8-bc80-ecd97b9622cc" Mar 13 09:31:27 crc kubenswrapper[4930]: E0313 09:31:27.820739 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f" Mar 13 09:31:27 crc kubenswrapper[4930]: E0313 09:31:27.820900 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pv4gt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-operator-controller-manager-bbc5b68f9-f45t2_openstack-operators(6c666666-f9ed-4183-9a92-e0334239aa3e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:27 crc kubenswrapper[4930]: E0313 09:31:27.822078 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" podUID="6c666666-f9ed-4183-9a92-e0334239aa3e" Mar 13 09:31:28 crc kubenswrapper[4930]: E0313 09:31:28.213424 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/ovn-operator@sha256:2f63ddf5c95c6c82f6e04bc9f7f20d56dc003614647726ab00276239eec40b7f\\\"\"" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" podUID="6c666666-f9ed-4183-9a92-e0334239aa3e" Mar 13 09:31:28 crc kubenswrapper[4930]: E0313 09:31:28.213488 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/cinder-operator@sha256:7c0da25380c91ffd1940d75eaa71b6842a6a4cf4056e62d6b0d237897b74e4d9\\\"\"" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" podUID="fd7bda25-ce82-4cc8-bc80-ecd97b9622cc" Mar 13 09:31:31 crc kubenswrapper[4930]: E0313 09:31:31.229715 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721" Mar 13 09:31:31 crc kubenswrapper[4930]: E0313 09:31:31.229907 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jvbpx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-776c5696bf-545tf_openstack-operators(a28dc996-b000-4680-b9f1-d770e91f0cba): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:31 crc kubenswrapper[4930]: E0313 09:31:31.230977 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" podUID="a28dc996-b000-4680-b9f1-d770e91f0cba" Mar 13 09:31:31 crc kubenswrapper[4930]: E0313 09:31:31.892229 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571" Mar 13 09:31:31 crc kubenswrapper[4930]: E0313 09:31:31.892922 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-pfggs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod octavia-operator-controller-manager-5f4f55cb5c-qx24r_openstack-operators(561d468b-4138-4196-8652-38b9eb4519ea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:31 crc kubenswrapper[4930]: E0313 09:31:31.894371 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" podUID="561d468b-4138-4196-8652-38b9eb4519ea" Mar 13 09:31:32 crc kubenswrapper[4930]: E0313 09:31:32.242028 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/octavia-operator@sha256:18fe6f2f0be7e736db86ff2d600af12a753e14b0a03232ce4f03629a89905571\\\"\"" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" podUID="561d468b-4138-4196-8652-38b9eb4519ea" Mar 13 09:31:32 crc kubenswrapper[4930]: E0313 09:31:32.247865 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/neutron-operator@sha256:5fe5351a3de5e1267112d52cd81477a01d47f90be713cc5439c76543a4c33721\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" podUID="a28dc996-b000-4680-b9f1-d770e91f0cba" Mar 13 09:31:33 crc kubenswrapper[4930]: I0313 09:31:33.032617 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:33 crc kubenswrapper[4930]: I0313 09:31:33.032706 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:33 crc kubenswrapper[4930]: I0313 09:31:33.039596 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-metrics-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:33 crc kubenswrapper[4930]: I0313 09:31:33.043599 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/984e8970-7f0d-4ce8-87c2-56c787857225-webhook-certs\") pod \"openstack-operator-controller-manager-6d5df8c674-hkqw8\" (UID: \"984e8970-7f0d-4ce8-87c2-56c787857225\") " pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:33 crc kubenswrapper[4930]: I0313 09:31:33.245927 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:33 crc kubenswrapper[4930]: E0313 09:31:33.874580 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:d9bffb59bb7f9f0a6cb103c3986fd2c1bdb13ce6349c39427a690858cbd754d6" Mar 13 09:31:33 crc kubenswrapper[4930]: E0313 09:31:33.874867 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:d9bffb59bb7f9f0a6cb103c3986fd2c1bdb13ce6349c39427a690858cbd754d6,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dqhs8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-6d9d6b584d-xlwng_openstack-operators(12a9ad60-3dff-443d-9f09-b57c4a373aa2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:33 crc kubenswrapper[4930]: E0313 09:31:33.876131 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" podUID="12a9ad60-3dff-443d-9f09-b57c4a373aa2" Mar 13 09:31:34 crc kubenswrapper[4930]: E0313 09:31:34.256833 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:d9bffb59bb7f9f0a6cb103c3986fd2c1bdb13ce6349c39427a690858cbd754d6\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" podUID="12a9ad60-3dff-443d-9f09-b57c4a373aa2" Mar 13 09:31:36 crc kubenswrapper[4930]: E0313 09:31:36.408126 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.103:5001/openstack-k8s-operators/telemetry-operator:1a1a9a719889b8cdda26cbd675f0005643a8f9f2" Mar 13 09:31:36 crc kubenswrapper[4930]: E0313 09:31:36.408853 4930 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="38.102.83.103:5001/openstack-k8s-operators/telemetry-operator:1a1a9a719889b8cdda26cbd675f0005643a8f9f2" Mar 13 09:31:36 crc kubenswrapper[4930]: E0313 09:31:36.409098 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:38.102.83.103:5001/openstack-k8s-operators/telemetry-operator:1a1a9a719889b8cdda26cbd675f0005643a8f9f2,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-56c7q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-5ffc4556d7-9h5cs_openstack-operators(1a25b1bf-af64-4f8c-920c-c450ebc66b41): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:36 crc kubenswrapper[4930]: E0313 09:31:36.410412 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podUID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" Mar 13 09:31:36 crc kubenswrapper[4930]: I0313 09:31:36.942445 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5"] Mar 13 09:31:37 crc kubenswrapper[4930]: E0313 09:31:37.015059 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Mar 13 09:31:37 crc kubenswrapper[4930]: E0313 09:31:37.015232 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-nlbgn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-wdbjc_openstack-operators(eaa30848-0297-4cdc-8c48-379fe75ad943): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:31:37 crc kubenswrapper[4930]: E0313 09:31:37.016642 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" podUID="eaa30848-0297-4cdc-8c48-379fe75ad943" Mar 13 09:31:37 crc kubenswrapper[4930]: I0313 09:31:37.296098 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" event={"ID":"28acf0f4-52a0-40f4-9a05-abbdcc79fb33","Type":"ContainerStarted","Data":"a4853c59d2895063280bb5caa522b1debaab5e071f4ed7fb679901b6c03175fd"} Mar 13 09:31:37 crc kubenswrapper[4930]: I0313 09:31:37.478201 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw"] Mar 13 09:31:37 crc kubenswrapper[4930]: I0313 09:31:37.573016 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8"] Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.307391 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" event={"ID":"d2ea88fb-18e6-4297-bb18-48d0e1d566be","Type":"ContainerStarted","Data":"636f02decec7b2fe9fec3d5b5cf8b8827855866bf3b417c9eb69ae8682673d24"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.307755 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.310733 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" event={"ID":"c92ab5a7-1ad7-4584-9954-0047b879bc47","Type":"ContainerStarted","Data":"bef4938f9775237068dc2648d30bf2194f0e292f76ab09dd660e9ff1b6f3dfe7"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.311064 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.312300 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" event={"ID":"585f7875-1138-4fcb-b89a-0a5f208f04b7","Type":"ContainerStarted","Data":"f5f7c5707765273a3986d62ab759e5d84b2ba1c0237b05d4605c6be1db8124ef"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.312469 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.316982 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" event={"ID":"5ede266e-f821-4c72-a408-16b4b71a483e","Type":"ContainerStarted","Data":"66f655479be8666581f68e249831689cd2727cd6f0a84d5972378141ffd47b30"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.317193 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.318599 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" event={"ID":"2e04d1ac-077a-4dc5-91a4-c9b22c352ac2","Type":"ContainerStarted","Data":"16812e88101fbc43011a0e3cc1bddc4032f3210f3674087f01d47035ff1ca0f3"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.319334 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.328403 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" event={"ID":"7601eb51-d514-4a96-848c-af0eae32f54f","Type":"ContainerStarted","Data":"66b8998f3f3c80f7f39237e0a5d4e8e0a32b7418f3b00b083e4ddb92c182c187"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.329371 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.335404 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" event={"ID":"4a2edceb-a605-4edd-8905-e35177c281c7","Type":"ContainerStarted","Data":"23016c16ef1a5471a7505508fd9b3982cbb758a10918d5468e3aee6c23f2d5e1"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.337062 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.354604 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" event={"ID":"ee0102ec-b840-47fb-bae7-4b6e04c231fa","Type":"ContainerStarted","Data":"7e831e52c1819263bfa933af9b44c0f466c70b5937bd4fe68f7127e634f8371f"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.356766 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.380228 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" podStartSLOduration=4.95206951 podStartE2EDuration="38.380199864s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:01.75223281 +0000 UTC m=+1102.502147487" lastFinishedPulling="2026-03-13 09:31:35.180363164 +0000 UTC m=+1135.930277841" observedRunningTime="2026-03-13 09:31:38.344967007 +0000 UTC m=+1139.094881684" watchObservedRunningTime="2026-03-13 09:31:38.380199864 +0000 UTC m=+1139.130114541" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.403183 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" event={"ID":"d12dd258-eed7-41ff-b2e9-f124e189e9aa","Type":"ContainerStarted","Data":"c5cac048d12a7e021f57f4c12b739852202f280c019356bd27436508c7f570d6"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.404642 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" event={"ID":"0d606b70-8e59-4303-963c-54c0be0a3800","Type":"ContainerStarted","Data":"6476e8dbcb813e652fee013be9726c717cf430303fba9ea91c144a0a4be6940d"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.406021 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.413105 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" event={"ID":"43277746-d9dc-4452-bcaf-8c4668ebf473","Type":"ContainerStarted","Data":"1bb9a94359afd2c13fabb115aebf66419c0dd534113d473822cac1e6f8bbb166"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.414465 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.420744 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" event={"ID":"984e8970-7f0d-4ce8-87c2-56c787857225","Type":"ContainerStarted","Data":"edef4108e148a17d772fce891c0e8427fba0dc7b6348a77940cf86c36917742e"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.420824 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" event={"ID":"984e8970-7f0d-4ce8-87c2-56c787857225","Type":"ContainerStarted","Data":"16f8274544fd667800b754f9fa2b5434b14b805318cfc64d95149f880bcb2aee"} Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.420932 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.452413 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" podStartSLOduration=4.293440207 podStartE2EDuration="38.45239245s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.874639314 +0000 UTC m=+1103.624553991" lastFinishedPulling="2026-03-13 09:31:37.033591547 +0000 UTC m=+1137.783506234" observedRunningTime="2026-03-13 09:31:38.381360623 +0000 UTC m=+1139.131275300" watchObservedRunningTime="2026-03-13 09:31:38.45239245 +0000 UTC m=+1139.202307127" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.512779 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" podStartSLOduration=5.061077018 podStartE2EDuration="38.512754468s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:01.728690764 +0000 UTC m=+1102.478605441" lastFinishedPulling="2026-03-13 09:31:35.180368204 +0000 UTC m=+1135.930282891" observedRunningTime="2026-03-13 09:31:38.420283182 +0000 UTC m=+1139.170197879" watchObservedRunningTime="2026-03-13 09:31:38.512754468 +0000 UTC m=+1139.262669145" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.531756 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" podStartSLOduration=4.289960056 podStartE2EDuration="38.531732386s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.845006955 +0000 UTC m=+1103.594921632" lastFinishedPulling="2026-03-13 09:31:37.086779285 +0000 UTC m=+1137.836693962" observedRunningTime="2026-03-13 09:31:38.472191718 +0000 UTC m=+1139.222106395" watchObservedRunningTime="2026-03-13 09:31:38.531732386 +0000 UTC m=+1139.281647063" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.538104 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" podStartSLOduration=3.877451942 podStartE2EDuration="38.538082626s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.406469006 +0000 UTC m=+1103.156383683" lastFinishedPulling="2026-03-13 09:31:37.06709965 +0000 UTC m=+1137.817014367" observedRunningTime="2026-03-13 09:31:38.511078996 +0000 UTC m=+1139.260993673" watchObservedRunningTime="2026-03-13 09:31:38.538082626 +0000 UTC m=+1139.287997303" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.552775 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" podStartSLOduration=4.075288434 podStartE2EDuration="38.552759235s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.556098065 +0000 UTC m=+1103.306012742" lastFinishedPulling="2026-03-13 09:31:37.033568866 +0000 UTC m=+1137.783483543" observedRunningTime="2026-03-13 09:31:38.545784939 +0000 UTC m=+1139.295699616" watchObservedRunningTime="2026-03-13 09:31:38.552759235 +0000 UTC m=+1139.302673902" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.585561 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" podStartSLOduration=4.344417162 podStartE2EDuration="39.58554523s" podCreationTimestamp="2026-03-13 09:30:59 +0000 UTC" firstStartedPulling="2026-03-13 09:31:01.052594525 +0000 UTC m=+1101.802509202" lastFinishedPulling="2026-03-13 09:31:36.293722593 +0000 UTC m=+1137.043637270" observedRunningTime="2026-03-13 09:31:38.58040039 +0000 UTC m=+1139.330315067" watchObservedRunningTime="2026-03-13 09:31:38.58554523 +0000 UTC m=+1139.335459907" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.611549 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" podStartSLOduration=3.256421765 podStartE2EDuration="38.611530343s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:01.73775921 +0000 UTC m=+1102.487673887" lastFinishedPulling="2026-03-13 09:31:37.092867788 +0000 UTC m=+1137.842782465" observedRunningTime="2026-03-13 09:31:38.604739873 +0000 UTC m=+1139.354654550" watchObservedRunningTime="2026-03-13 09:31:38.611530343 +0000 UTC m=+1139.361445020" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.636779 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" podStartSLOduration=38.636755728 podStartE2EDuration="38.636755728s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:31:38.634258445 +0000 UTC m=+1139.384173132" watchObservedRunningTime="2026-03-13 09:31:38.636755728 +0000 UTC m=+1139.386670405" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.669267 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" podStartSLOduration=4.781144191 podStartE2EDuration="38.669247505s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.556132896 +0000 UTC m=+1103.306047573" lastFinishedPulling="2026-03-13 09:31:36.4442362 +0000 UTC m=+1137.194150887" observedRunningTime="2026-03-13 09:31:38.660379962 +0000 UTC m=+1139.410294639" watchObservedRunningTime="2026-03-13 09:31:38.669247505 +0000 UTC m=+1139.419162182" Mar 13 09:31:38 crc kubenswrapper[4930]: I0313 09:31:38.693229 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" podStartSLOduration=6.312413113 podStartE2EDuration="38.693208748s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.903048342 +0000 UTC m=+1103.652963019" lastFinishedPulling="2026-03-13 09:31:35.283843937 +0000 UTC m=+1136.033758654" observedRunningTime="2026-03-13 09:31:38.687820383 +0000 UTC m=+1139.437735060" watchObservedRunningTime="2026-03-13 09:31:38.693208748 +0000 UTC m=+1139.443123425" Mar 13 09:31:39 crc kubenswrapper[4930]: I0313 09:31:39.429319 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" event={"ID":"4051662d-ae7a-4f38-9aa5-5eecab9693f3","Type":"ContainerStarted","Data":"ed62b13cfd28c5f4a4a4fa9a940b2c319695c76af11cc94c3d55861434a47b97"} Mar 13 09:31:39 crc kubenswrapper[4930]: I0313 09:31:39.430610 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" Mar 13 09:31:39 crc kubenswrapper[4930]: I0313 09:31:39.448328 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" podStartSLOduration=3.059687196 podStartE2EDuration="39.448306355s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:01.490412736 +0000 UTC m=+1102.240327413" lastFinishedPulling="2026-03-13 09:31:37.879031895 +0000 UTC m=+1138.628946572" observedRunningTime="2026-03-13 09:31:39.4457239 +0000 UTC m=+1140.195638587" watchObservedRunningTime="2026-03-13 09:31:39.448306355 +0000 UTC m=+1140.198221032" Mar 13 09:31:40 crc kubenswrapper[4930]: I0313 09:31:40.437700 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" event={"ID":"fd7bda25-ce82-4cc8-bc80-ecd97b9622cc","Type":"ContainerStarted","Data":"b7e3725e6882cd818f7397953267d9578e6b05b941ba8d004a00cfb1e03c0e93"} Mar 13 09:31:40 crc kubenswrapper[4930]: I0313 09:31:40.438272 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" Mar 13 09:31:40 crc kubenswrapper[4930]: I0313 09:31:40.439197 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" event={"ID":"046b75aa-f3a0-467d-8624-6a82aa748b32","Type":"ContainerStarted","Data":"3dda63aa85c8aed23d3f18135a76880272ed2f16e9b5fa1aa7705eb5b0fcf843"} Mar 13 09:31:40 crc kubenswrapper[4930]: I0313 09:31:40.439676 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" Mar 13 09:31:40 crc kubenswrapper[4930]: I0313 09:31:40.458227 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" podStartSLOduration=3.421111476 podStartE2EDuration="41.458209672s" podCreationTimestamp="2026-03-13 09:30:59 +0000 UTC" firstStartedPulling="2026-03-13 09:31:01.464734596 +0000 UTC m=+1102.214649273" lastFinishedPulling="2026-03-13 09:31:39.501832792 +0000 UTC m=+1140.251747469" observedRunningTime="2026-03-13 09:31:40.452052787 +0000 UTC m=+1141.201967484" watchObservedRunningTime="2026-03-13 09:31:40.458209672 +0000 UTC m=+1141.208124349" Mar 13 09:31:40 crc kubenswrapper[4930]: I0313 09:31:40.472223 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" podStartSLOduration=3.840861192 podStartE2EDuration="40.472203904s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.873175877 +0000 UTC m=+1103.623090554" lastFinishedPulling="2026-03-13 09:31:39.504518589 +0000 UTC m=+1140.254433266" observedRunningTime="2026-03-13 09:31:40.465887465 +0000 UTC m=+1141.215802162" watchObservedRunningTime="2026-03-13 09:31:40.472203904 +0000 UTC m=+1141.222118581" Mar 13 09:31:42 crc kubenswrapper[4930]: I0313 09:31:42.308765 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:31:42 crc kubenswrapper[4930]: I0313 09:31:42.310199 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:31:42 crc kubenswrapper[4930]: I0313 09:31:42.453866 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" event={"ID":"28acf0f4-52a0-40f4-9a05-abbdcc79fb33","Type":"ContainerStarted","Data":"c43349f90ba1b8393ad9053e8f986ad79cb593de5bbb86e03065a7dc3fb0773e"} Mar 13 09:31:42 crc kubenswrapper[4930]: I0313 09:31:42.454221 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:42 crc kubenswrapper[4930]: I0313 09:31:42.455575 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" event={"ID":"d12dd258-eed7-41ff-b2e9-f124e189e9aa","Type":"ContainerStarted","Data":"75930369908ef1072aaf7fe4e5761c5ec63a3683818121a47bffb7adfd8d5149"} Mar 13 09:31:42 crc kubenswrapper[4930]: I0313 09:31:42.455717 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:42 crc kubenswrapper[4930]: I0313 09:31:42.478920 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" podStartSLOduration=37.892800452 podStartE2EDuration="42.478903147s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:37.044939572 +0000 UTC m=+1137.794854249" lastFinishedPulling="2026-03-13 09:31:41.631042267 +0000 UTC m=+1142.380956944" observedRunningTime="2026-03-13 09:31:42.477332888 +0000 UTC m=+1143.227247555" watchObservedRunningTime="2026-03-13 09:31:42.478903147 +0000 UTC m=+1143.228817824" Mar 13 09:31:42 crc kubenswrapper[4930]: I0313 09:31:42.500982 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" podStartSLOduration=38.372803687 podStartE2EDuration="42.500967312s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:37.502453051 +0000 UTC m=+1138.252367728" lastFinishedPulling="2026-03-13 09:31:41.630616676 +0000 UTC m=+1142.380531353" observedRunningTime="2026-03-13 09:31:42.499596898 +0000 UTC m=+1143.249511575" watchObservedRunningTime="2026-03-13 09:31:42.500967312 +0000 UTC m=+1143.250881989" Mar 13 09:31:43 crc kubenswrapper[4930]: I0313 09:31:43.252994 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 09:31:43 crc kubenswrapper[4930]: I0313 09:31:43.464035 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" event={"ID":"47a82845-cf9f-4fc8-a35f-4548a9c7127e","Type":"ContainerStarted","Data":"b3f2db2d703fcbd1b9e8e356fa8567251136b01277698cdafa5e474fbfad9220"} Mar 13 09:31:43 crc kubenswrapper[4930]: I0313 09:31:43.464488 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" Mar 13 09:31:43 crc kubenswrapper[4930]: I0313 09:31:43.489690 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" podStartSLOduration=3.457687702 podStartE2EDuration="43.489670476s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.533878421 +0000 UTC m=+1103.283793088" lastFinishedPulling="2026-03-13 09:31:42.565861185 +0000 UTC m=+1143.315775862" observedRunningTime="2026-03-13 09:31:43.482228109 +0000 UTC m=+1144.232142786" watchObservedRunningTime="2026-03-13 09:31:43.489670476 +0000 UTC m=+1144.239585153" Mar 13 09:31:44 crc kubenswrapper[4930]: I0313 09:31:44.473936 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" event={"ID":"6c666666-f9ed-4183-9a92-e0334239aa3e","Type":"ContainerStarted","Data":"dfce868c7dce10d78c2a69dc4ef53eff1b3a6d17a6918af786fa4f7f7a2fc4a5"} Mar 13 09:31:44 crc kubenswrapper[4930]: I0313 09:31:44.474711 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" Mar 13 09:31:44 crc kubenswrapper[4930]: I0313 09:31:44.494163 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" podStartSLOduration=3.34670669 podStartE2EDuration="44.494146926s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.533699857 +0000 UTC m=+1103.283614534" lastFinishedPulling="2026-03-13 09:31:43.681140093 +0000 UTC m=+1144.431054770" observedRunningTime="2026-03-13 09:31:44.492808013 +0000 UTC m=+1145.242722680" watchObservedRunningTime="2026-03-13 09:31:44.494146926 +0000 UTC m=+1145.244061603" Mar 13 09:31:45 crc kubenswrapper[4930]: I0313 09:31:45.482013 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" event={"ID":"561d468b-4138-4196-8652-38b9eb4519ea","Type":"ContainerStarted","Data":"085319f76f22396e728776931f44e84c2719ef4e001d4ee30bd83176e2397167"} Mar 13 09:31:45 crc kubenswrapper[4930]: I0313 09:31:45.482473 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" Mar 13 09:31:45 crc kubenswrapper[4930]: I0313 09:31:45.484815 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" event={"ID":"a28dc996-b000-4680-b9f1-d770e91f0cba","Type":"ContainerStarted","Data":"8fe24ba25156609c8f02fdd769b945deaf2a3c827ce6ef3473034cf5411efc6a"} Mar 13 09:31:45 crc kubenswrapper[4930]: I0313 09:31:45.485026 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" Mar 13 09:31:45 crc kubenswrapper[4930]: I0313 09:31:45.500450 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" podStartSLOduration=3.552308951 podStartE2EDuration="45.500420901s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.53344665 +0000 UTC m=+1103.283361327" lastFinishedPulling="2026-03-13 09:31:44.4815586 +0000 UTC m=+1145.231473277" observedRunningTime="2026-03-13 09:31:45.49361266 +0000 UTC m=+1146.243527337" watchObservedRunningTime="2026-03-13 09:31:45.500420901 +0000 UTC m=+1146.250335578" Mar 13 09:31:45 crc kubenswrapper[4930]: I0313 09:31:45.515410 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" podStartSLOduration=3.135438205 podStartE2EDuration="45.515391308s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.53584802 +0000 UTC m=+1103.285762697" lastFinishedPulling="2026-03-13 09:31:44.915801123 +0000 UTC m=+1145.665715800" observedRunningTime="2026-03-13 09:31:45.506766981 +0000 UTC m=+1146.256681658" watchObservedRunningTime="2026-03-13 09:31:45.515391308 +0000 UTC m=+1146.265305985" Mar 13 09:31:46 crc kubenswrapper[4930]: I0313 09:31:46.440675 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 09:31:46 crc kubenswrapper[4930]: I0313 09:31:46.453565 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 09:31:47 crc kubenswrapper[4930]: I0313 09:31:47.504873 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" event={"ID":"12a9ad60-3dff-443d-9f09-b57c4a373aa2","Type":"ContainerStarted","Data":"ce5c76e24c78752c9b7167b4cfbd24d21047fa57c49a42b56abf2eb7928ddf77"} Mar 13 09:31:47 crc kubenswrapper[4930]: I0313 09:31:47.506220 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" Mar 13 09:31:47 crc kubenswrapper[4930]: I0313 09:31:47.526144 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" podStartSLOduration=3.389071215 podStartE2EDuration="47.526121743s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.53624128 +0000 UTC m=+1103.286155957" lastFinishedPulling="2026-03-13 09:31:46.673291798 +0000 UTC m=+1147.423206485" observedRunningTime="2026-03-13 09:31:47.524701458 +0000 UTC m=+1148.274616145" watchObservedRunningTime="2026-03-13 09:31:47.526121743 +0000 UTC m=+1148.276036430" Mar 13 09:31:49 crc kubenswrapper[4930]: E0313 09:31:49.971951 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" podUID="eaa30848-0297-4cdc-8c48-379fe75ad943" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.340468 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.347769 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.403052 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.436285 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.460129 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.589005 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.630393 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.654191 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.686131 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.715893 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.735759 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.807212 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.866652 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" Mar 13 09:31:50 crc kubenswrapper[4930]: I0313 09:31:50.878789 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" Mar 13 09:31:51 crc kubenswrapper[4930]: I0313 09:31:51.194884 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" Mar 13 09:31:51 crc kubenswrapper[4930]: I0313 09:31:51.281214 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" Mar 13 09:31:51 crc kubenswrapper[4930]: I0313 09:31:51.355913 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" Mar 13 09:31:51 crc kubenswrapper[4930]: E0313 09:31:51.974141 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"38.102.83.103:5001/openstack-k8s-operators/telemetry-operator:1a1a9a719889b8cdda26cbd675f0005643a8f9f2\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podUID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" Mar 13 09:32:00 crc kubenswrapper[4930]: I0313 09:32:00.143478 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556572-l9zlc"] Mar 13 09:32:00 crc kubenswrapper[4930]: I0313 09:32:00.145414 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556572-l9zlc" Mar 13 09:32:00 crc kubenswrapper[4930]: I0313 09:32:00.147618 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:32:00 crc kubenswrapper[4930]: I0313 09:32:00.148070 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:32:00 crc kubenswrapper[4930]: I0313 09:32:00.151676 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:32:00 crc kubenswrapper[4930]: I0313 09:32:00.152272 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556572-l9zlc"] Mar 13 09:32:00 crc kubenswrapper[4930]: I0313 09:32:00.311750 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llg96\" (UniqueName: \"kubernetes.io/projected/27777941-a761-4e19-b4bf-6228b4e0221c-kube-api-access-llg96\") pod \"auto-csr-approver-29556572-l9zlc\" (UID: \"27777941-a761-4e19-b4bf-6228b4e0221c\") " pod="openshift-infra/auto-csr-approver-29556572-l9zlc" Mar 13 09:32:00 crc kubenswrapper[4930]: I0313 09:32:00.414043 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llg96\" (UniqueName: \"kubernetes.io/projected/27777941-a761-4e19-b4bf-6228b4e0221c-kube-api-access-llg96\") pod \"auto-csr-approver-29556572-l9zlc\" (UID: \"27777941-a761-4e19-b4bf-6228b4e0221c\") " pod="openshift-infra/auto-csr-approver-29556572-l9zlc" Mar 13 09:32:00 crc kubenswrapper[4930]: I0313 09:32:00.437002 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llg96\" (UniqueName: \"kubernetes.io/projected/27777941-a761-4e19-b4bf-6228b4e0221c-kube-api-access-llg96\") pod \"auto-csr-approver-29556572-l9zlc\" (UID: \"27777941-a761-4e19-b4bf-6228b4e0221c\") " pod="openshift-infra/auto-csr-approver-29556572-l9zlc" Mar 13 09:32:00 crc kubenswrapper[4930]: I0313 09:32:00.473477 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556572-l9zlc" Mar 13 09:32:00 crc kubenswrapper[4930]: I0313 09:32:00.516863 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" Mar 13 09:32:00 crc kubenswrapper[4930]: I0313 09:32:00.938263 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556572-l9zlc"] Mar 13 09:32:00 crc kubenswrapper[4930]: W0313 09:32:00.939623 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27777941_a761_4e19_b4bf_6228b4e0221c.slice/crio-48030f4ff8b3b646feb2a2b99d1977f7cec991d497a2d0729f727a22f4929a9c WatchSource:0}: Error finding container 48030f4ff8b3b646feb2a2b99d1977f7cec991d497a2d0729f727a22f4929a9c: Status 404 returned error can't find the container with id 48030f4ff8b3b646feb2a2b99d1977f7cec991d497a2d0729f727a22f4929a9c Mar 13 09:32:01 crc kubenswrapper[4930]: I0313 09:32:01.618938 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556572-l9zlc" event={"ID":"27777941-a761-4e19-b4bf-6228b4e0221c","Type":"ContainerStarted","Data":"48030f4ff8b3b646feb2a2b99d1977f7cec991d497a2d0729f727a22f4929a9c"} Mar 13 09:32:02 crc kubenswrapper[4930]: I0313 09:32:02.627839 4930 generic.go:334] "Generic (PLEG): container finished" podID="27777941-a761-4e19-b4bf-6228b4e0221c" containerID="4abda406b7dbafb34ba62ab222a8d16f0082647d4e4d291a18bc2c69deca6f09" exitCode=0 Mar 13 09:32:02 crc kubenswrapper[4930]: I0313 09:32:02.627938 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556572-l9zlc" event={"ID":"27777941-a761-4e19-b4bf-6228b4e0221c","Type":"ContainerDied","Data":"4abda406b7dbafb34ba62ab222a8d16f0082647d4e4d291a18bc2c69deca6f09"} Mar 13 09:32:02 crc kubenswrapper[4930]: I0313 09:32:02.629721 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" event={"ID":"eaa30848-0297-4cdc-8c48-379fe75ad943","Type":"ContainerStarted","Data":"23e03e25065102c20cceab910acdf822979ed2c10766d9f61d048f5a67b22d19"} Mar 13 09:32:02 crc kubenswrapper[4930]: I0313 09:32:02.665387 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-wdbjc" podStartSLOduration=4.110695012 podStartE2EDuration="1m2.66536626s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.947644403 +0000 UTC m=+1103.697559080" lastFinishedPulling="2026-03-13 09:32:01.502315651 +0000 UTC m=+1162.252230328" observedRunningTime="2026-03-13 09:32:02.654550498 +0000 UTC m=+1163.404465175" watchObservedRunningTime="2026-03-13 09:32:02.66536626 +0000 UTC m=+1163.415280937" Mar 13 09:32:03 crc kubenswrapper[4930]: I0313 09:32:03.638354 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" event={"ID":"1a25b1bf-af64-4f8c-920c-c450ebc66b41","Type":"ContainerStarted","Data":"8d450e36e2a6be6462e86bc3137c2d91e9b9dd1006f05f4fb76254cf36f5c093"} Mar 13 09:32:03 crc kubenswrapper[4930]: I0313 09:32:03.638902 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" Mar 13 09:32:03 crc kubenswrapper[4930]: I0313 09:32:03.655021 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podStartSLOduration=3.538030313 podStartE2EDuration="1m3.655004417s" podCreationTimestamp="2026-03-13 09:31:00 +0000 UTC" firstStartedPulling="2026-03-13 09:31:02.919571234 +0000 UTC m=+1103.669485911" lastFinishedPulling="2026-03-13 09:32:03.036545338 +0000 UTC m=+1163.786460015" observedRunningTime="2026-03-13 09:32:03.653507399 +0000 UTC m=+1164.403422076" watchObservedRunningTime="2026-03-13 09:32:03.655004417 +0000 UTC m=+1164.404919094" Mar 13 09:32:03 crc kubenswrapper[4930]: I0313 09:32:03.964755 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556572-l9zlc" Mar 13 09:32:04 crc kubenswrapper[4930]: I0313 09:32:04.096740 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llg96\" (UniqueName: \"kubernetes.io/projected/27777941-a761-4e19-b4bf-6228b4e0221c-kube-api-access-llg96\") pod \"27777941-a761-4e19-b4bf-6228b4e0221c\" (UID: \"27777941-a761-4e19-b4bf-6228b4e0221c\") " Mar 13 09:32:04 crc kubenswrapper[4930]: I0313 09:32:04.105700 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27777941-a761-4e19-b4bf-6228b4e0221c-kube-api-access-llg96" (OuterVolumeSpecName: "kube-api-access-llg96") pod "27777941-a761-4e19-b4bf-6228b4e0221c" (UID: "27777941-a761-4e19-b4bf-6228b4e0221c"). InnerVolumeSpecName "kube-api-access-llg96". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:32:04 crc kubenswrapper[4930]: I0313 09:32:04.198922 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llg96\" (UniqueName: \"kubernetes.io/projected/27777941-a761-4e19-b4bf-6228b4e0221c-kube-api-access-llg96\") on node \"crc\" DevicePath \"\"" Mar 13 09:32:04 crc kubenswrapper[4930]: I0313 09:32:04.647833 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556572-l9zlc" Mar 13 09:32:04 crc kubenswrapper[4930]: I0313 09:32:04.647819 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556572-l9zlc" event={"ID":"27777941-a761-4e19-b4bf-6228b4e0221c","Type":"ContainerDied","Data":"48030f4ff8b3b646feb2a2b99d1977f7cec991d497a2d0729f727a22f4929a9c"} Mar 13 09:32:04 crc kubenswrapper[4930]: I0313 09:32:04.647885 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="48030f4ff8b3b646feb2a2b99d1977f7cec991d497a2d0729f727a22f4929a9c" Mar 13 09:32:05 crc kubenswrapper[4930]: I0313 09:32:05.031705 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556566-7lxzt"] Mar 13 09:32:05 crc kubenswrapper[4930]: I0313 09:32:05.040914 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556566-7lxzt"] Mar 13 09:32:05 crc kubenswrapper[4930]: I0313 09:32:05.982730 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a5d1519-2396-4ba8-983c-189a8ca117c3" path="/var/lib/kubelet/pods/2a5d1519-2396-4ba8-983c-189a8ca117c3/volumes" Mar 13 09:32:11 crc kubenswrapper[4930]: I0313 09:32:11.228986 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" Mar 13 09:32:12 crc kubenswrapper[4930]: I0313 09:32:12.308469 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:32:12 crc kubenswrapper[4930]: I0313 09:32:12.308840 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:32:12 crc kubenswrapper[4930]: I0313 09:32:12.308887 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:32:12 crc kubenswrapper[4930]: I0313 09:32:12.309589 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c06f278e05cf4208d0a63eab8c9e20f18523a288ed0cb0082dfaac4bd5791ab7"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 09:32:12 crc kubenswrapper[4930]: I0313 09:32:12.309640 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://c06f278e05cf4208d0a63eab8c9e20f18523a288ed0cb0082dfaac4bd5791ab7" gracePeriod=600 Mar 13 09:32:12 crc kubenswrapper[4930]: I0313 09:32:12.715875 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="c06f278e05cf4208d0a63eab8c9e20f18523a288ed0cb0082dfaac4bd5791ab7" exitCode=0 Mar 13 09:32:12 crc kubenswrapper[4930]: I0313 09:32:12.715944 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"c06f278e05cf4208d0a63eab8c9e20f18523a288ed0cb0082dfaac4bd5791ab7"} Mar 13 09:32:12 crc kubenswrapper[4930]: I0313 09:32:12.716273 4930 scope.go:117] "RemoveContainer" containerID="2a734be71e61d016548cb1b09f6ff35d98f17d0ce5f81f146f9d61495866a932" Mar 13 09:32:13 crc kubenswrapper[4930]: I0313 09:32:13.725419 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"9a600e0833fe6da30e3cb3376c93b675518bc4f00e36348b0e36bfb8a45c09d7"} Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.335248 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s8cx2"] Mar 13 09:32:28 crc kubenswrapper[4930]: E0313 09:32:28.336262 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27777941-a761-4e19-b4bf-6228b4e0221c" containerName="oc" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.336278 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="27777941-a761-4e19-b4bf-6228b4e0221c" containerName="oc" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.336483 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="27777941-a761-4e19-b4bf-6228b4e0221c" containerName="oc" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.343520 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.349586 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.349649 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.349649 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.349846 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-29cn5" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.366521 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s8cx2"] Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.390627 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e70e48b-fe5b-4292-b609-baec526562df-config\") pod \"dnsmasq-dns-675f4bcbfc-s8cx2\" (UID: \"8e70e48b-fe5b-4292-b609-baec526562df\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.390717 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxnq7\" (UniqueName: \"kubernetes.io/projected/8e70e48b-fe5b-4292-b609-baec526562df-kube-api-access-nxnq7\") pod \"dnsmasq-dns-675f4bcbfc-s8cx2\" (UID: \"8e70e48b-fe5b-4292-b609-baec526562df\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.492279 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e70e48b-fe5b-4292-b609-baec526562df-config\") pod \"dnsmasq-dns-675f4bcbfc-s8cx2\" (UID: \"8e70e48b-fe5b-4292-b609-baec526562df\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.492348 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxnq7\" (UniqueName: \"kubernetes.io/projected/8e70e48b-fe5b-4292-b609-baec526562df-kube-api-access-nxnq7\") pod \"dnsmasq-dns-675f4bcbfc-s8cx2\" (UID: \"8e70e48b-fe5b-4292-b609-baec526562df\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.493625 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e70e48b-fe5b-4292-b609-baec526562df-config\") pod \"dnsmasq-dns-675f4bcbfc-s8cx2\" (UID: \"8e70e48b-fe5b-4292-b609-baec526562df\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.537518 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxnq7\" (UniqueName: \"kubernetes.io/projected/8e70e48b-fe5b-4292-b609-baec526562df-kube-api-access-nxnq7\") pod \"dnsmasq-dns-675f4bcbfc-s8cx2\" (UID: \"8e70e48b-fe5b-4292-b609-baec526562df\") " pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.590638 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5dkxq"] Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.592312 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.596054 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.599081 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5dkxq"] Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.661080 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.694801 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-5dkxq\" (UID: \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.694889 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-config\") pod \"dnsmasq-dns-78dd6ddcc-5dkxq\" (UID: \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.694959 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jpk2q\" (UniqueName: \"kubernetes.io/projected/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-kube-api-access-jpk2q\") pod \"dnsmasq-dns-78dd6ddcc-5dkxq\" (UID: \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.798503 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jpk2q\" (UniqueName: \"kubernetes.io/projected/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-kube-api-access-jpk2q\") pod \"dnsmasq-dns-78dd6ddcc-5dkxq\" (UID: \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.798825 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-5dkxq\" (UID: \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.798881 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-config\") pod \"dnsmasq-dns-78dd6ddcc-5dkxq\" (UID: \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.800114 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-config\") pod \"dnsmasq-dns-78dd6ddcc-5dkxq\" (UID: \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.800136 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-5dkxq\" (UID: \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.830225 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jpk2q\" (UniqueName: \"kubernetes.io/projected/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-kube-api-access-jpk2q\") pod \"dnsmasq-dns-78dd6ddcc-5dkxq\" (UID: \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\") " pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:32:28 crc kubenswrapper[4930]: I0313 09:32:28.909053 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:32:29 crc kubenswrapper[4930]: I0313 09:32:29.155706 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s8cx2"] Mar 13 09:32:29 crc kubenswrapper[4930]: I0313 09:32:29.351390 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5dkxq"] Mar 13 09:32:29 crc kubenswrapper[4930]: I0313 09:32:29.901311 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" event={"ID":"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e","Type":"ContainerStarted","Data":"a8c9aa9547826c1e1014f4184fa4e3fa5ff6628f97f9c7f9260cb4ce4b2f32ec"} Mar 13 09:32:29 crc kubenswrapper[4930]: I0313 09:32:29.903592 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" event={"ID":"8e70e48b-fe5b-4292-b609-baec526562df","Type":"ContainerStarted","Data":"a65c626c011c6638c7eda1d99ea332b14e5811e4120bd36c6c07eb81695a6de8"} Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.042451 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s8cx2"] Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.079973 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-zc722"] Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.081707 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.103851 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-zc722"] Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.146885 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b302d040-3268-4633-b651-560d5a0645d2-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-zc722\" (UID: \"b302d040-3268-4633-b651-560d5a0645d2\") " pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.146940 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qjdhf\" (UniqueName: \"kubernetes.io/projected/b302d040-3268-4633-b651-560d5a0645d2-kube-api-access-qjdhf\") pod \"dnsmasq-dns-5ccc8479f9-zc722\" (UID: \"b302d040-3268-4633-b651-560d5a0645d2\") " pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.146986 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b302d040-3268-4633-b651-560d5a0645d2-config\") pod \"dnsmasq-dns-5ccc8479f9-zc722\" (UID: \"b302d040-3268-4633-b651-560d5a0645d2\") " pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.248628 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b302d040-3268-4633-b651-560d5a0645d2-config\") pod \"dnsmasq-dns-5ccc8479f9-zc722\" (UID: \"b302d040-3268-4633-b651-560d5a0645d2\") " pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.248772 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b302d040-3268-4633-b651-560d5a0645d2-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-zc722\" (UID: \"b302d040-3268-4633-b651-560d5a0645d2\") " pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.248800 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qjdhf\" (UniqueName: \"kubernetes.io/projected/b302d040-3268-4633-b651-560d5a0645d2-kube-api-access-qjdhf\") pod \"dnsmasq-dns-5ccc8479f9-zc722\" (UID: \"b302d040-3268-4633-b651-560d5a0645d2\") " pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.249768 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b302d040-3268-4633-b651-560d5a0645d2-config\") pod \"dnsmasq-dns-5ccc8479f9-zc722\" (UID: \"b302d040-3268-4633-b651-560d5a0645d2\") " pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.250487 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b302d040-3268-4633-b651-560d5a0645d2-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-zc722\" (UID: \"b302d040-3268-4633-b651-560d5a0645d2\") " pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.292319 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qjdhf\" (UniqueName: \"kubernetes.io/projected/b302d040-3268-4633-b651-560d5a0645d2-kube-api-access-qjdhf\") pod \"dnsmasq-dns-5ccc8479f9-zc722\" (UID: \"b302d040-3268-4633-b651-560d5a0645d2\") " pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.390031 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5dkxq"] Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.441670 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.443801 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zsn7s"] Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.453672 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.468513 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zsn7s"] Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.556175 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba559492-1275-4e57-ba68-b94af80e931b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zsn7s\" (UID: \"ba559492-1275-4e57-ba68-b94af80e931b\") " pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.556238 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9hwlr\" (UniqueName: \"kubernetes.io/projected/ba559492-1275-4e57-ba68-b94af80e931b-kube-api-access-9hwlr\") pod \"dnsmasq-dns-57d769cc4f-zsn7s\" (UID: \"ba559492-1275-4e57-ba68-b94af80e931b\") " pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.556333 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba559492-1275-4e57-ba68-b94af80e931b-config\") pod \"dnsmasq-dns-57d769cc4f-zsn7s\" (UID: \"ba559492-1275-4e57-ba68-b94af80e931b\") " pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.661347 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba559492-1275-4e57-ba68-b94af80e931b-config\") pod \"dnsmasq-dns-57d769cc4f-zsn7s\" (UID: \"ba559492-1275-4e57-ba68-b94af80e931b\") " pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.661428 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba559492-1275-4e57-ba68-b94af80e931b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zsn7s\" (UID: \"ba559492-1275-4e57-ba68-b94af80e931b\") " pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.661478 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9hwlr\" (UniqueName: \"kubernetes.io/projected/ba559492-1275-4e57-ba68-b94af80e931b-kube-api-access-9hwlr\") pod \"dnsmasq-dns-57d769cc4f-zsn7s\" (UID: \"ba559492-1275-4e57-ba68-b94af80e931b\") " pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.662360 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba559492-1275-4e57-ba68-b94af80e931b-config\") pod \"dnsmasq-dns-57d769cc4f-zsn7s\" (UID: \"ba559492-1275-4e57-ba68-b94af80e931b\") " pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.662401 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba559492-1275-4e57-ba68-b94af80e931b-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-zsn7s\" (UID: \"ba559492-1275-4e57-ba68-b94af80e931b\") " pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.712097 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9hwlr\" (UniqueName: \"kubernetes.io/projected/ba559492-1275-4e57-ba68-b94af80e931b-kube-api-access-9hwlr\") pod \"dnsmasq-dns-57d769cc4f-zsn7s\" (UID: \"ba559492-1275-4e57-ba68-b94af80e931b\") " pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:32:31 crc kubenswrapper[4930]: I0313 09:32:31.822478 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.198223 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-zc722"] Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.212481 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.243184 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.248660 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.254831 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.255284 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.254933 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.255063 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.255716 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.255062 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.255465 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-h595l" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.296299 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.296372 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.296399 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.296549 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.296626 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.296652 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.296674 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-cb39e555-44ae-4231-8aab-8695177f64b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.296808 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.296868 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65qcg\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-kube-api-access-65qcg\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.296904 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.296942 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.398645 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.399956 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.400013 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.400051 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.400069 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.400089 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-cb39e555-44ae-4231-8aab-8695177f64b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.400131 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.400154 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65qcg\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-kube-api-access-65qcg\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.400172 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.400196 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.400240 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.403489 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.404062 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.405818 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.406086 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.406789 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.407543 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.407583 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-cb39e555-44ae-4231-8aab-8695177f64b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/31037259f8435172bc4f2806d5e4fc542b3e0cf5ddb9acb377b0574429b2c6a4/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.409501 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.410006 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.410869 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.417834 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65qcg\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-kube-api-access-65qcg\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.422563 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.487526 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-cb39e555-44ae-4231-8aab-8695177f64b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") pod \"rabbitmq-cell1-server-0\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.590661 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.687585 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.689552 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.692058 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.692816 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.693029 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.693191 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-ljjf2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.693333 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.693644 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.693820 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.697000 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.751723 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-2"] Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.758972 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.785533 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-1"] Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.787855 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.805162 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809005 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-server-conf\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809055 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809080 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809099 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/897aa260-40c2-42d2-b59f-964a1a40116c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809127 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809156 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809190 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jcf7\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-kube-api-access-2jcf7\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809215 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809233 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809260 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809275 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809298 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809322 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f69fa32-5b72-4f9b-9176-d43ee011605d-pod-info\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809350 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-config-data\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809393 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-config-data\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809406 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-server-conf\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809422 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809459 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/897aa260-40c2-42d2-b59f-964a1a40116c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809497 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809660 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809693 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6623f09-f961-411a-9323-8cbf953c7148-pod-info\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809715 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f69fa32-5b72-4f9b-9176-d43ee011605d-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809742 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-config-data\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809777 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6623f09-f961-411a-9323-8cbf953c7148-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809794 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809831 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809858 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dchqq\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-kube-api-access-dchqq\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809888 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809904 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809922 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809944 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809961 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4976\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-kube-api-access-j4976\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.809989 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.834586 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.911457 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6623f09-f961-411a-9323-8cbf953c7148-pod-info\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.911494 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f69fa32-5b72-4f9b-9176-d43ee011605d-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.911521 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-config-data\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.911547 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6623f09-f961-411a-9323-8cbf953c7148-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.911568 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.914391 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-config-data\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916168 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916534 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916594 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dchqq\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-kube-api-access-dchqq\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916637 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916656 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916683 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916708 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916745 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4976\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-kube-api-access-j4976\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916784 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916833 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-server-conf\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916863 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916884 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916904 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/897aa260-40c2-42d2-b59f-964a1a40116c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916957 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.916985 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.917409 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.918586 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.919676 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.922257 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.923864 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.925168 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6623f09-f961-411a-9323-8cbf953c7148-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.925804 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.926328 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zsn7s"] Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.926456 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jcf7\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-kube-api-access-2jcf7\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.926878 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6623f09-f961-411a-9323-8cbf953c7148-pod-info\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.926912 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.928137 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-server-conf\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.928368 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.928393 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/58fe0acdb728d3cb19cbd87a16786e051248a03f1d289ab57726095b513a2e26/globalmount\"" pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.928665 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.928704 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e16cd85ffa615f778b152995784bb8923bd8ede16bc037cea68a4cfdb6f64ffd/globalmount\"" pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.928737 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.928764 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/973977d077043ef8fddec8a941dafd7e9bdafd407131659801d4835fea31a670/globalmount\"" pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.929081 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.930512 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.930542 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.930572 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.930600 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f69fa32-5b72-4f9b-9176-d43ee011605d-pod-info\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.930639 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-config-data\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.930690 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-config-data\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.930715 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.930729 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-server-conf\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.930753 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/897aa260-40c2-42d2-b59f-964a1a40116c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.930799 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.930820 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.930962 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-server-conf\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.932309 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-server-conf\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.933629 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.933927 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-config-data\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.935320 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.936304 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.937689 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-config-data\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.941017 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f69fa32-5b72-4f9b-9176-d43ee011605d-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.941506 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/897aa260-40c2-42d2-b59f-964a1a40116c-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.944283 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f69fa32-5b72-4f9b-9176-d43ee011605d-pod-info\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.944480 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.945831 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jcf7\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-kube-api-access-2jcf7\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.947830 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.950082 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.950366 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dchqq\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-kube-api-access-dchqq\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.950894 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4976\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-kube-api-access-j4976\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.952282 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.966920 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/897aa260-40c2-42d2-b59f-964a1a40116c-pod-info\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.967343 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.968937 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" event={"ID":"b302d040-3268-4633-b651-560d5a0645d2","Type":"ContainerStarted","Data":"1f528838203ea0e6aa144f33f506efef885dd25f696e5faa83b463524e812027"} Mar 13 09:32:32 crc kubenswrapper[4930]: I0313 09:32:32.976205 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" event={"ID":"ba559492-1275-4e57-ba68-b94af80e931b","Type":"ContainerStarted","Data":"54cf96892290030dd9da19e9a5ae367c5d0b7159c943e58ab5de9b12c026fa36"} Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.011524 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\") pod \"rabbitmq-server-1\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " pod="openstack/rabbitmq-server-1" Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.015028 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\") pod \"rabbitmq-server-0\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " pod="openstack/rabbitmq-server-0" Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.020786 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.027183 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\") pod \"rabbitmq-server-2\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " pod="openstack/rabbitmq-server-2" Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.118903 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.159070 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.247735 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 09:32:33 crc kubenswrapper[4930]: W0313 09:32:33.261488 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddc44fa4e_c7fd_4a4b_a03c_cb02e3b47354.slice/crio-d9d4aecc2b7d2feebb05a1139df043bc7f059ac862bb780cb0ef82fafe45a341 WatchSource:0}: Error finding container d9d4aecc2b7d2feebb05a1139df043bc7f059ac862bb780cb0ef82fafe45a341: Status 404 returned error can't find the container with id d9d4aecc2b7d2feebb05a1139df043bc7f059ac862bb780cb0ef82fafe45a341 Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.575439 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 09:32:33 crc kubenswrapper[4930]: W0313 09:32:33.583692 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod897aa260_40c2_42d2_b59f_964a1a40116c.slice/crio-c4bd8903e11193a779d6a36c39193e29bcf85e6394c93049499223a6dcf37350 WatchSource:0}: Error finding container c4bd8903e11193a779d6a36c39193e29bcf85e6394c93049499223a6dcf37350: Status 404 returned error can't find the container with id c4bd8903e11193a779d6a36c39193e29bcf85e6394c93049499223a6dcf37350 Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.832536 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.846626 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 13 09:32:33 crc kubenswrapper[4930]: W0313 09:32:33.862737 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2f69fa32_5b72_4f9b_9176_d43ee011605d.slice/crio-63f2bb7080932271c381d7af1c9b19e5a24762dd7d2af49248b5c79d9709c96b WatchSource:0}: Error finding container 63f2bb7080932271c381d7af1c9b19e5a24762dd7d2af49248b5c79d9709c96b: Status 404 returned error can't find the container with id 63f2bb7080932271c381d7af1c9b19e5a24762dd7d2af49248b5c79d9709c96b Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.905242 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.907396 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.937686 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.938751 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.959902 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.959904 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.960317 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-djc7b" Mar 13 09:32:33 crc kubenswrapper[4930]: I0313 09:32:33.967869 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.009819 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354","Type":"ContainerStarted","Data":"d9d4aecc2b7d2feebb05a1139df043bc7f059ac862bb780cb0ef82fafe45a341"} Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.013812 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"897aa260-40c2-42d2-b59f-964a1a40116c","Type":"ContainerStarted","Data":"c4bd8903e11193a779d6a36c39193e29bcf85e6394c93049499223a6dcf37350"} Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.019005 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"2f69fa32-5b72-4f9b-9176-d43ee011605d","Type":"ContainerStarted","Data":"63f2bb7080932271c381d7af1c9b19e5a24762dd7d2af49248b5c79d9709c96b"} Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.025479 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a6623f09-f961-411a-9323-8cbf953c7148","Type":"ContainerStarted","Data":"2cedef14877fa4198d43afb9b9dc678dbb52d67071539ef6175ee15d5d9be9a1"} Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.080396 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-config-data-default\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.080503 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.080738 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-06bd7732-2282-4469-ae90-eb9278527ac5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-06bd7732-2282-4469-ae90-eb9278527ac5\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.080788 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpvkw\" (UniqueName: \"kubernetes.io/projected/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-kube-api-access-cpvkw\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.080845 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.080927 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-kolla-config\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.080953 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.081077 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.183003 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-config-data-default\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.183149 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.183255 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-06bd7732-2282-4469-ae90-eb9278527ac5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-06bd7732-2282-4469-ae90-eb9278527ac5\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.183693 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpvkw\" (UniqueName: \"kubernetes.io/projected/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-kube-api-access-cpvkw\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.183748 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.183788 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-kolla-config\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.183807 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.183837 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.184559 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-config-data-default\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.185533 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-config-data-generated\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.185599 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-kolla-config\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.186829 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-operator-scripts\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.191520 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.191556 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-06bd7732-2282-4469-ae90-eb9278527ac5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-06bd7732-2282-4469-ae90-eb9278527ac5\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/8107fcf42a5eee7178ecba2c94769d64ac1a900aab0a118ee625546b06bff0ed/globalmount\"" pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.201127 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.203753 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpvkw\" (UniqueName: \"kubernetes.io/projected/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-kube-api-access-cpvkw\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.215002 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/8adee7e3-81a7-4dcf-ad19-d50e1900d3e0-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.246002 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-06bd7732-2282-4469-ae90-eb9278527ac5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-06bd7732-2282-4469-ae90-eb9278527ac5\") pod \"openstack-galera-0\" (UID: \"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0\") " pod="openstack/openstack-galera-0" Mar 13 09:32:34 crc kubenswrapper[4930]: I0313 09:32:34.293868 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.054542 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 13 09:32:35 crc kubenswrapper[4930]: W0313 09:32:35.129350 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8adee7e3_81a7_4dcf_ad19_d50e1900d3e0.slice/crio-2bbab19c8945cf867ec201c1120e1ccc385ffef0d945875af8bdcf080cd23607 WatchSource:0}: Error finding container 2bbab19c8945cf867ec201c1120e1ccc385ffef0d945875af8bdcf080cd23607: Status 404 returned error can't find the container with id 2bbab19c8945cf867ec201c1120e1ccc385ffef0d945875af8bdcf080cd23607 Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.227171 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.229376 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.231978 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.232645 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-ndz8l" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.232971 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.233291 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.254148 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.311920 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fce00700-2a2f-4fd3-b165-2f041bc5105f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fce00700-2a2f-4fd3-b165-2f041bc5105f\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.311981 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3459cebe-06a7-428e-93ef-419677fbcb58-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.312010 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3459cebe-06a7-428e-93ef-419677fbcb58-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.312065 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3459cebe-06a7-428e-93ef-419677fbcb58-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.312086 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3459cebe-06a7-428e-93ef-419677fbcb58-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.312119 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vl4w\" (UniqueName: \"kubernetes.io/projected/3459cebe-06a7-428e-93ef-419677fbcb58-kube-api-access-9vl4w\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.312147 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3459cebe-06a7-428e-93ef-419677fbcb58-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.312180 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3459cebe-06a7-428e-93ef-419677fbcb58-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.414806 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3459cebe-06a7-428e-93ef-419677fbcb58-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.414854 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3459cebe-06a7-428e-93ef-419677fbcb58-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.414886 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vl4w\" (UniqueName: \"kubernetes.io/projected/3459cebe-06a7-428e-93ef-419677fbcb58-kube-api-access-9vl4w\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.414915 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3459cebe-06a7-428e-93ef-419677fbcb58-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.414952 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3459cebe-06a7-428e-93ef-419677fbcb58-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.415006 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fce00700-2a2f-4fd3-b165-2f041bc5105f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fce00700-2a2f-4fd3-b165-2f041bc5105f\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.415033 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3459cebe-06a7-428e-93ef-419677fbcb58-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.415056 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3459cebe-06a7-428e-93ef-419677fbcb58-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.415708 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/3459cebe-06a7-428e-93ef-419677fbcb58-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.416997 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3459cebe-06a7-428e-93ef-419677fbcb58-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.416997 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/3459cebe-06a7-428e-93ef-419677fbcb58-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.417954 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/3459cebe-06a7-428e-93ef-419677fbcb58-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.421365 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.421405 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fce00700-2a2f-4fd3-b165-2f041bc5105f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fce00700-2a2f-4fd3-b165-2f041bc5105f\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/4616cf682bbcd33605baa4439cf9e587700b84ad8e58789cd5fa063d1ac160c7/globalmount\"" pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.426246 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/3459cebe-06a7-428e-93ef-419677fbcb58-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.426259 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3459cebe-06a7-428e-93ef-419677fbcb58-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.437030 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vl4w\" (UniqueName: \"kubernetes.io/projected/3459cebe-06a7-428e-93ef-419677fbcb58-kube-api-access-9vl4w\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.497638 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fce00700-2a2f-4fd3-b165-2f041bc5105f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fce00700-2a2f-4fd3-b165-2f041bc5105f\") pod \"openstack-cell1-galera-0\" (UID: \"3459cebe-06a7-428e-93ef-419677fbcb58\") " pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.559969 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.564771 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.565957 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.571909 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.572100 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.572344 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-wtbn9" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.590994 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.721048 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-kolla-config\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.721140 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.721179 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgsmz\" (UniqueName: \"kubernetes.io/projected/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-kube-api-access-mgsmz\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.721222 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.721481 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-config-data\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.824401 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-kolla-config\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.824469 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.824525 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgsmz\" (UniqueName: \"kubernetes.io/projected/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-kube-api-access-mgsmz\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.824597 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.824671 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-config-data\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.825825 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-config-data\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.826251 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-kolla-config\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.832901 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.842067 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.845740 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgsmz\" (UniqueName: \"kubernetes.io/projected/9b328c28-0be8-4f5c-8e76-bf33c8ee1726-kube-api-access-mgsmz\") pod \"memcached-0\" (UID: \"9b328c28-0be8-4f5c-8e76-bf33c8ee1726\") " pod="openstack/memcached-0" Mar 13 09:32:35 crc kubenswrapper[4930]: I0313 09:32:35.921722 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 13 09:32:36 crc kubenswrapper[4930]: I0313 09:32:36.114447 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0","Type":"ContainerStarted","Data":"2bbab19c8945cf867ec201c1120e1ccc385ffef0d945875af8bdcf080cd23607"} Mar 13 09:32:36 crc kubenswrapper[4930]: I0313 09:32:36.404273 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 13 09:32:36 crc kubenswrapper[4930]: I0313 09:32:36.582024 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 13 09:32:37 crc kubenswrapper[4930]: I0313 09:32:37.183874 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9b328c28-0be8-4f5c-8e76-bf33c8ee1726","Type":"ContainerStarted","Data":"397d6b76eda0c2668aaee8d2c62b763a241e2fc11c5a7604e07ba1f1bd238598"} Mar 13 09:32:37 crc kubenswrapper[4930]: I0313 09:32:37.218602 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3459cebe-06a7-428e-93ef-419677fbcb58","Type":"ContainerStarted","Data":"4b5ad8a99c8693b49a9a39611993e9fa987721ecb03048235e84fcce02520682"} Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.215002 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.216825 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.230841 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-p85mk" Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.248813 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.309594 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v565g\" (UniqueName: \"kubernetes.io/projected/ecb88f7d-f623-4395-9c32-566a3ef1f0c9-kube-api-access-v565g\") pod \"kube-state-metrics-0\" (UID: \"ecb88f7d-f623-4395-9c32-566a3ef1f0c9\") " pod="openstack/kube-state-metrics-0" Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.422566 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v565g\" (UniqueName: \"kubernetes.io/projected/ecb88f7d-f623-4395-9c32-566a3ef1f0c9-kube-api-access-v565g\") pod \"kube-state-metrics-0\" (UID: \"ecb88f7d-f623-4395-9c32-566a3ef1f0c9\") " pod="openstack/kube-state-metrics-0" Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.508378 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v565g\" (UniqueName: \"kubernetes.io/projected/ecb88f7d-f623-4395-9c32-566a3ef1f0c9-kube-api-access-v565g\") pod \"kube-state-metrics-0\" (UID: \"ecb88f7d-f623-4395-9c32-566a3ef1f0c9\") " pod="openstack/kube-state-metrics-0" Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.592163 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.972135 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8"] Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.974059 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8" Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.978694 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.978917 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-xdkp8" Mar 13 09:32:38 crc kubenswrapper[4930]: I0313 09:32:38.989648 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8"] Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.037671 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w5k9\" (UniqueName: \"kubernetes.io/projected/5e86c9f9-c149-46a6-9210-ec642593320d-kube-api-access-9w5k9\") pod \"observability-ui-dashboards-66cbf594b5-kpvf8\" (UID: \"5e86c9f9-c149-46a6-9210-ec642593320d\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.038078 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e86c9f9-c149-46a6-9210-ec642593320d-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-kpvf8\" (UID: \"5e86c9f9-c149-46a6-9210-ec642593320d\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.139812 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e86c9f9-c149-46a6-9210-ec642593320d-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-kpvf8\" (UID: \"5e86c9f9-c149-46a6-9210-ec642593320d\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.139924 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w5k9\" (UniqueName: \"kubernetes.io/projected/5e86c9f9-c149-46a6-9210-ec642593320d-kube-api-access-9w5k9\") pod \"observability-ui-dashboards-66cbf594b5-kpvf8\" (UID: \"5e86c9f9-c149-46a6-9210-ec642593320d\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8" Mar 13 09:32:39 crc kubenswrapper[4930]: E0313 09:32:39.139961 4930 secret.go:188] Couldn't get secret openshift-operators/observability-ui-dashboards: secret "observability-ui-dashboards" not found Mar 13 09:32:39 crc kubenswrapper[4930]: E0313 09:32:39.140088 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5e86c9f9-c149-46a6-9210-ec642593320d-serving-cert podName:5e86c9f9-c149-46a6-9210-ec642593320d nodeName:}" failed. No retries permitted until 2026-03-13 09:32:39.640037295 +0000 UTC m=+1200.389951972 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/5e86c9f9-c149-46a6-9210-ec642593320d-serving-cert") pod "observability-ui-dashboards-66cbf594b5-kpvf8" (UID: "5e86c9f9-c149-46a6-9210-ec642593320d") : secret "observability-ui-dashboards" not found Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.253656 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w5k9\" (UniqueName: \"kubernetes.io/projected/5e86c9f9-c149-46a6-9210-ec642593320d-kube-api-access-9w5k9\") pod \"observability-ui-dashboards-66cbf594b5-kpvf8\" (UID: \"5e86c9f9-c149-46a6-9210-ec642593320d\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.395764 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.401118 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.422073 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.430719 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.430917 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.431071 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.431180 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.431270 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.431445 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-gsfzj" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.431570 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.444380 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.500666 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-7656bb8967-5vbb4"] Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.501959 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.550899 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7656bb8967-5vbb4"] Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.589911 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/91a9c833-bbb1-4341-899b-d922b4066dc0-oauth-serving-cert\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.589984 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590014 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhs7j\" (UniqueName: \"kubernetes.io/projected/91a9c833-bbb1-4341-899b-d922b4066dc0-kube-api-access-mhs7j\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590036 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/91a9c833-bbb1-4341-899b-d922b4066dc0-console-config\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590085 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/91a9c833-bbb1-4341-899b-d922b4066dc0-console-oauth-config\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590115 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/267a25c3-20d9-4b6d-b097-d14b9534faa1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590134 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mvlv\" (UniqueName: \"kubernetes.io/projected/267a25c3-20d9-4b6d-b097-d14b9534faa1-kube-api-access-9mvlv\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590155 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-config\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590177 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/267a25c3-20d9-4b6d-b097-d14b9534faa1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590206 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590226 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/91a9c833-bbb1-4341-899b-d922b4066dc0-console-serving-cert\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590258 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91a9c833-bbb1-4341-899b-d922b4066dc0-trusted-ca-bundle\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590287 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590311 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/91a9c833-bbb1-4341-899b-d922b4066dc0-service-ca\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590341 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590358 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.590393 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696516 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696574 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696621 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696665 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/91a9c833-bbb1-4341-899b-d922b4066dc0-oauth-serving-cert\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696705 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696727 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhs7j\" (UniqueName: \"kubernetes.io/projected/91a9c833-bbb1-4341-899b-d922b4066dc0-kube-api-access-mhs7j\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696745 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/91a9c833-bbb1-4341-899b-d922b4066dc0-console-config\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696769 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e86c9f9-c149-46a6-9210-ec642593320d-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-kpvf8\" (UID: \"5e86c9f9-c149-46a6-9210-ec642593320d\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696790 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/91a9c833-bbb1-4341-899b-d922b4066dc0-console-oauth-config\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696817 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/267a25c3-20d9-4b6d-b097-d14b9534faa1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696833 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mvlv\" (UniqueName: \"kubernetes.io/projected/267a25c3-20d9-4b6d-b097-d14b9534faa1-kube-api-access-9mvlv\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696854 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-config\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696872 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/267a25c3-20d9-4b6d-b097-d14b9534faa1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696897 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696916 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/91a9c833-bbb1-4341-899b-d922b4066dc0-console-serving-cert\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696941 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91a9c833-bbb1-4341-899b-d922b4066dc0-trusted-ca-bundle\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696963 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.696986 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/91a9c833-bbb1-4341-899b-d922b4066dc0-service-ca\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.698019 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/91a9c833-bbb1-4341-899b-d922b4066dc0-service-ca\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.708163 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/267a25c3-20d9-4b6d-b097-d14b9534faa1-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.710093 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.712404 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/91a9c833-bbb1-4341-899b-d922b4066dc0-console-oauth-config\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.712770 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/91a9c833-bbb1-4341-899b-d922b4066dc0-console-config\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.713392 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.714140 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/91a9c833-bbb1-4341-899b-d922b4066dc0-trusted-ca-bundle\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.714684 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.715102 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/91a9c833-bbb1-4341-899b-d922b4066dc0-oauth-serving-cert\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.729688 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.729739 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e182641cca8f9c7abeba89f962dbe513ebbaf8cfc329d127147730d16d8b04fe/globalmount\"" pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.730028 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/267a25c3-20d9-4b6d-b097-d14b9534faa1-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.734476 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/91a9c833-bbb1-4341-899b-d922b4066dc0-console-serving-cert\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.734590 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.739837 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mvlv\" (UniqueName: \"kubernetes.io/projected/267a25c3-20d9-4b6d-b097-d14b9534faa1-kube-api-access-9mvlv\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.740448 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.743029 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e86c9f9-c149-46a6-9210-ec642593320d-serving-cert\") pod \"observability-ui-dashboards-66cbf594b5-kpvf8\" (UID: \"5e86c9f9-c149-46a6-9210-ec642593320d\") " pod="openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.748532 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-config\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.781960 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhs7j\" (UniqueName: \"kubernetes.io/projected/91a9c833-bbb1-4341-899b-d922b4066dc0-kube-api-access-mhs7j\") pod \"console-7656bb8967-5vbb4\" (UID: \"91a9c833-bbb1-4341-899b-d922b4066dc0\") " pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.835447 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.881205 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\") pod \"prometheus-metric-storage-0\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:39 crc kubenswrapper[4930]: I0313 09:32:39.913741 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8" Mar 13 09:32:40 crc kubenswrapper[4930]: I0313 09:32:40.069267 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-gsfzj" Mar 13 09:32:40 crc kubenswrapper[4930]: I0313 09:32:40.074952 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 13 09:32:40 crc kubenswrapper[4930]: I0313 09:32:40.202014 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.207710 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.210378 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.216308 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.216990 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.217169 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.217343 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.217717 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-scqhr" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.228121 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.365335 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e3a3580-5ee8-4853-8b81-43768ea76d20-config\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.365408 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e3a3580-5ee8-4853-8b81-43768ea76d20-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.365485 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e3a3580-5ee8-4853-8b81-43768ea76d20-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.365506 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e3a3580-5ee8-4853-8b81-43768ea76d20-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.365532 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dt6qh\" (UniqueName: \"kubernetes.io/projected/6e3a3580-5ee8-4853-8b81-43768ea76d20-kube-api-access-dt6qh\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.365561 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-09aceef7-51a3-4ab2-96b7-9cabad96574b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-09aceef7-51a3-4ab2-96b7-9cabad96574b\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.365591 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e3a3580-5ee8-4853-8b81-43768ea76d20-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.365622 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e3a3580-5ee8-4853-8b81-43768ea76d20-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.466974 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e3a3580-5ee8-4853-8b81-43768ea76d20-config\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.467028 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e3a3580-5ee8-4853-8b81-43768ea76d20-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.467085 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e3a3580-5ee8-4853-8b81-43768ea76d20-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.467104 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e3a3580-5ee8-4853-8b81-43768ea76d20-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.467133 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dt6qh\" (UniqueName: \"kubernetes.io/projected/6e3a3580-5ee8-4853-8b81-43768ea76d20-kube-api-access-dt6qh\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.467167 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-09aceef7-51a3-4ab2-96b7-9cabad96574b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-09aceef7-51a3-4ab2-96b7-9cabad96574b\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.467201 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e3a3580-5ee8-4853-8b81-43768ea76d20-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.467236 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e3a3580-5ee8-4853-8b81-43768ea76d20-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.470015 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6e3a3580-5ee8-4853-8b81-43768ea76d20-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.470763 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6e3a3580-5ee8-4853-8b81-43768ea76d20-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.470820 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6e3a3580-5ee8-4853-8b81-43768ea76d20-config\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.480147 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6e3a3580-5ee8-4853-8b81-43768ea76d20-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.492638 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e3a3580-5ee8-4853-8b81-43768ea76d20-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.493217 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6e3a3580-5ee8-4853-8b81-43768ea76d20-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.500204 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.500257 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-09aceef7-51a3-4ab2-96b7-9cabad96574b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-09aceef7-51a3-4ab2-96b7-9cabad96574b\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/75f52a9d604a11c32ecf59fb616a6c9e10bf8ea62469c9c98d5e6be6b506d3d9/globalmount\"" pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.526709 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dt6qh\" (UniqueName: \"kubernetes.io/projected/6e3a3580-5ee8-4853-8b81-43768ea76d20-kube-api-access-dt6qh\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.567323 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-09aceef7-51a3-4ab2-96b7-9cabad96574b\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-09aceef7-51a3-4ab2-96b7-9cabad96574b\") pod \"ovsdbserver-nb-0\" (UID: \"6e3a3580-5ee8-4853-8b81-43768ea76d20\") " pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.675908 4930 scope.go:117] "RemoveContainer" containerID="502126d38b429293d46a422c3fa582f3f7ef0f53f85780f5c35628e0822fe318" Mar 13 09:32:41 crc kubenswrapper[4930]: I0313 09:32:41.844635 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.666667 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fghdj"] Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.669005 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.687098 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.687479 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.688192 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-z2d6l" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.701041 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fghdj"] Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.738567 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-zw9cp"] Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.740815 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.759736 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-zw9cp"] Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.796464 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2307735f-0b44-41ed-a622-74d62c05810c-var-run\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.796586 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2307735f-0b44-41ed-a622-74d62c05810c-scripts\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.796648 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2307735f-0b44-41ed-a622-74d62c05810c-var-run-ovn\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.796725 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2307735f-0b44-41ed-a622-74d62c05810c-combined-ca-bundle\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.796859 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2307735f-0b44-41ed-a622-74d62c05810c-ovn-controller-tls-certs\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.797073 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2d5r\" (UniqueName: \"kubernetes.io/projected/2307735f-0b44-41ed-a622-74d62c05810c-kube-api-access-d2d5r\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.797213 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2307735f-0b44-41ed-a622-74d62c05810c-var-log-ovn\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899160 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2307735f-0b44-41ed-a622-74d62c05810c-ovn-controller-tls-certs\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899224 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/20dc8601-3918-4d9c-993e-5ec14c0140b4-var-lib\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899259 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2d5r\" (UniqueName: \"kubernetes.io/projected/2307735f-0b44-41ed-a622-74d62c05810c-kube-api-access-d2d5r\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899287 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2307735f-0b44-41ed-a622-74d62c05810c-var-log-ovn\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899345 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20dc8601-3918-4d9c-993e-5ec14c0140b4-scripts\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899388 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2307735f-0b44-41ed-a622-74d62c05810c-var-run\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899412 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2307735f-0b44-41ed-a622-74d62c05810c-scripts\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899426 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5ws5\" (UniqueName: \"kubernetes.io/projected/20dc8601-3918-4d9c-993e-5ec14c0140b4-kube-api-access-x5ws5\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899463 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2307735f-0b44-41ed-a622-74d62c05810c-var-run-ovn\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899480 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20dc8601-3918-4d9c-993e-5ec14c0140b4-var-run\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899510 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/20dc8601-3918-4d9c-993e-5ec14c0140b4-var-log\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899525 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/20dc8601-3918-4d9c-993e-5ec14c0140b4-etc-ovs\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.899545 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2307735f-0b44-41ed-a622-74d62c05810c-combined-ca-bundle\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.903487 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2307735f-0b44-41ed-a622-74d62c05810c-var-log-ovn\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.903966 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2307735f-0b44-41ed-a622-74d62c05810c-var-run-ovn\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.904153 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2307735f-0b44-41ed-a622-74d62c05810c-var-run\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.907190 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2307735f-0b44-41ed-a622-74d62c05810c-combined-ca-bundle\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.907706 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2307735f-0b44-41ed-a622-74d62c05810c-scripts\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.909265 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2307735f-0b44-41ed-a622-74d62c05810c-ovn-controller-tls-certs\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:42 crc kubenswrapper[4930]: I0313 09:32:42.919177 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2d5r\" (UniqueName: \"kubernetes.io/projected/2307735f-0b44-41ed-a622-74d62c05810c-kube-api-access-d2d5r\") pod \"ovn-controller-fghdj\" (UID: \"2307735f-0b44-41ed-a622-74d62c05810c\") " pod="openstack/ovn-controller-fghdj" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.002386 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/20dc8601-3918-4d9c-993e-5ec14c0140b4-var-lib\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.002755 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20dc8601-3918-4d9c-993e-5ec14c0140b4-scripts\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.002808 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5ws5\" (UniqueName: \"kubernetes.io/projected/20dc8601-3918-4d9c-993e-5ec14c0140b4-kube-api-access-x5ws5\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.002831 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20dc8601-3918-4d9c-993e-5ec14c0140b4-var-run\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.002871 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/20dc8601-3918-4d9c-993e-5ec14c0140b4-var-log\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.002886 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/20dc8601-3918-4d9c-993e-5ec14c0140b4-etc-ovs\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.003110 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/20dc8601-3918-4d9c-993e-5ec14c0140b4-etc-ovs\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.003227 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/20dc8601-3918-4d9c-993e-5ec14c0140b4-var-lib\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.004576 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/20dc8601-3918-4d9c-993e-5ec14c0140b4-var-run\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.004878 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/20dc8601-3918-4d9c-993e-5ec14c0140b4-var-log\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.005067 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/20dc8601-3918-4d9c-993e-5ec14c0140b4-scripts\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.015208 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.021977 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5ws5\" (UniqueName: \"kubernetes.io/projected/20dc8601-3918-4d9c-993e-5ec14c0140b4-kube-api-access-x5ws5\") pod \"ovn-controller-ovs-zw9cp\" (UID: \"20dc8601-3918-4d9c-993e-5ec14c0140b4\") " pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:43 crc kubenswrapper[4930]: I0313 09:32:43.062499 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.623305 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.628181 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.634051 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.634281 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.634500 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-n2t48" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.634661 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.637696 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.766580 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcdcr\" (UniqueName: \"kubernetes.io/projected/d08733c1-5776-499e-8dbe-1cc29dd9716e-kube-api-access-bcdcr\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.766652 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-500b5fba-67f0-43bb-9fb7-814dd209dfe8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-500b5fba-67f0-43bb-9fb7-814dd209dfe8\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.766704 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d08733c1-5776-499e-8dbe-1cc29dd9716e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.766771 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d08733c1-5776-499e-8dbe-1cc29dd9716e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.766814 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d08733c1-5776-499e-8dbe-1cc29dd9716e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.766830 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d08733c1-5776-499e-8dbe-1cc29dd9716e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.766854 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d08733c1-5776-499e-8dbe-1cc29dd9716e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.766868 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d08733c1-5776-499e-8dbe-1cc29dd9716e-config\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.868595 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d08733c1-5776-499e-8dbe-1cc29dd9716e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.868636 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d08733c1-5776-499e-8dbe-1cc29dd9716e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.868665 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d08733c1-5776-499e-8dbe-1cc29dd9716e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.868681 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d08733c1-5776-499e-8dbe-1cc29dd9716e-config\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.868759 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcdcr\" (UniqueName: \"kubernetes.io/projected/d08733c1-5776-499e-8dbe-1cc29dd9716e-kube-api-access-bcdcr\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.868798 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-500b5fba-67f0-43bb-9fb7-814dd209dfe8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-500b5fba-67f0-43bb-9fb7-814dd209dfe8\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.868830 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d08733c1-5776-499e-8dbe-1cc29dd9716e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.868896 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d08733c1-5776-499e-8dbe-1cc29dd9716e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.874034 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d08733c1-5776-499e-8dbe-1cc29dd9716e-config\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.874458 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/d08733c1-5776-499e-8dbe-1cc29dd9716e-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.875787 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d08733c1-5776-499e-8dbe-1cc29dd9716e-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.876860 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/d08733c1-5776-499e-8dbe-1cc29dd9716e-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.876915 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/d08733c1-5776-499e-8dbe-1cc29dd9716e-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.878793 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.878828 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-500b5fba-67f0-43bb-9fb7-814dd209dfe8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-500b5fba-67f0-43bb-9fb7-814dd209dfe8\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/77161475fdfd8bf065c7b1e128a8afd33154cf56943bbf0b7317d2ff9608a3f7/globalmount\"" pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.886148 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d08733c1-5776-499e-8dbe-1cc29dd9716e-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.894289 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcdcr\" (UniqueName: \"kubernetes.io/projected/d08733c1-5776-499e-8dbe-1cc29dd9716e-kube-api-access-bcdcr\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.910959 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-500b5fba-67f0-43bb-9fb7-814dd209dfe8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-500b5fba-67f0-43bb-9fb7-814dd209dfe8\") pod \"ovsdbserver-sb-0\" (UID: \"d08733c1-5776-499e-8dbe-1cc29dd9716e\") " pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:44 crc kubenswrapper[4930]: I0313 09:32:44.971388 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 13 09:32:47 crc kubenswrapper[4930]: I0313 09:32:47.506223 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ecb88f7d-f623-4395-9c32-566a3ef1f0c9","Type":"ContainerStarted","Data":"f88eee0f524c5469e2db7689bf53be0693d5d333b8f7d542ad8c5a61cf5572b6"} Mar 13 09:32:53 crc kubenswrapper[4930]: E0313 09:32:53.818885 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Mar 13 09:32:53 crc kubenswrapper[4930]: E0313 09:32:53.820575 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j4976,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-2_openstack(a6623f09-f961-411a-9323-8cbf953c7148): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:32:53 crc kubenswrapper[4930]: E0313 09:32:53.821836 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-2" podUID="a6623f09-f961-411a-9323-8cbf953c7148" Mar 13 09:32:54 crc kubenswrapper[4930]: E0313 09:32:54.592103 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-2" podUID="a6623f09-f961-411a-9323-8cbf953c7148" Mar 13 09:32:57 crc kubenswrapper[4930]: E0313 09:32:57.316829 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached:current-podified" Mar 13 09:32:57 crc kubenswrapper[4930]: E0313 09:32:57.317496 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:n7ch55h65bh5cdh59fhdch78h55bh54fh54h5b6h5c8h54h55fh99h5cch5c6h549h5cch64chcdh5c6h94h5bdh698h5bch58bh5dch97hb7hf4h64bq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/certs/memcached.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:memcached-tls-certs,ReadOnly:true,MountPath:/var/lib/config-data/tls/private/memcached.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-mgsmz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_openstack(9b328c28-0be8-4f5c-8e76-bf33c8ee1726): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:32:57 crc kubenswrapper[4930]: E0313 09:32:57.319478 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/memcached-0" podUID="9b328c28-0be8-4f5c-8e76-bf33c8ee1726" Mar 13 09:32:57 crc kubenswrapper[4930]: E0313 09:32:57.626956 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached:current-podified\\\"\"" pod="openstack/memcached-0" podUID="9b328c28-0be8-4f5c-8e76-bf33c8ee1726" Mar 13 09:33:02 crc kubenswrapper[4930]: I0313 09:33:02.640536 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-7656bb8967-5vbb4"] Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.251603 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.252040 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nxnq7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-s8cx2_openstack(8e70e48b-fe5b-4292-b609-baec526562df): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.254744 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" podUID="8e70e48b-fe5b-4292-b609-baec526562df" Mar 13 09:33:04 crc kubenswrapper[4930]: W0313 09:33:04.289946 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91a9c833_bbb1_4341_899b_d922b4066dc0.slice/crio-354b14b8eaa2c9d9e785b01929882977fee04e52587da3b0cb69f200d1c5c4a0 WatchSource:0}: Error finding container 354b14b8eaa2c9d9e785b01929882977fee04e52587da3b0cb69f200d1c5c4a0: Status 404 returned error can't find the container with id 354b14b8eaa2c9d9e785b01929882977fee04e52587da3b0cb69f200d1c5c4a0 Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.305868 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.306000 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qjdhf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-zc722_openstack(b302d040-3268-4633-b651-560d5a0645d2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.307318 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" podUID="b302d040-3268-4633-b651-560d5a0645d2" Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.325460 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.325586 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-jpk2q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-5dkxq_openstack(92bf50f8-1f71-4cf9-9f36-7209ab28dc5e): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.326836 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" podUID="92bf50f8-1f71-4cf9-9f36-7209ab28dc5e" Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.379092 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.379240 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9hwlr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-zsn7s_openstack(ba559492-1275-4e57-ba68-b94af80e931b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.382086 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" podUID="ba559492-1275-4e57-ba68-b94af80e931b" Mar 13 09:33:04 crc kubenswrapper[4930]: I0313 09:33:04.699317 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7656bb8967-5vbb4" event={"ID":"91a9c833-bbb1-4341-899b-d922b4066dc0","Type":"ContainerStarted","Data":"354b14b8eaa2c9d9e785b01929882977fee04e52587da3b0cb69f200d1c5c4a0"} Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.710807 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" podUID="ba559492-1275-4e57-ba68-b94af80e931b" Mar 13 09:33:04 crc kubenswrapper[4930]: E0313 09:33:04.717931 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" podUID="b302d040-3268-4633-b651-560d5a0645d2" Mar 13 09:33:04 crc kubenswrapper[4930]: I0313 09:33:04.767484 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fghdj"] Mar 13 09:33:04 crc kubenswrapper[4930]: I0313 09:33:04.951679 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.280610 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8"] Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.343269 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-zw9cp"] Mar 13 09:33:05 crc kubenswrapper[4930]: W0313 09:33:05.416509 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e86c9f9_c149_46a6_9210_ec642593320d.slice/crio-d3285e071a18504d4d870a3faaf96160e76a7255e61e4db20d4cc95d660d599b WatchSource:0}: Error finding container d3285e071a18504d4d870a3faaf96160e76a7255e61e4db20d4cc95d660d599b: Status 404 returned error can't find the container with id d3285e071a18504d4d870a3faaf96160e76a7255e61e4db20d4cc95d660d599b Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.518908 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.544549 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.712016 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj" event={"ID":"2307735f-0b44-41ed-a622-74d62c05810c","Type":"ContainerStarted","Data":"5068f2a863f7e4a203e82491e45ed5759e9a5539315972d4a6129437385acffb"} Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.713387 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8" event={"ID":"5e86c9f9-c149-46a6-9210-ec642593320d","Type":"ContainerStarted","Data":"d3285e071a18504d4d870a3faaf96160e76a7255e61e4db20d4cc95d660d599b"} Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.713950 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-config\") pod \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\" (UID: \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\") " Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.714175 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jpk2q\" (UniqueName: \"kubernetes.io/projected/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-kube-api-access-jpk2q\") pod \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\" (UID: \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\") " Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.714376 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nxnq7\" (UniqueName: \"kubernetes.io/projected/8e70e48b-fe5b-4292-b609-baec526562df-kube-api-access-nxnq7\") pod \"8e70e48b-fe5b-4292-b609-baec526562df\" (UID: \"8e70e48b-fe5b-4292-b609-baec526562df\") " Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.714415 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-dns-svc\") pod \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\" (UID: \"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e\") " Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.714674 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e70e48b-fe5b-4292-b609-baec526562df-config\") pod \"8e70e48b-fe5b-4292-b609-baec526562df\" (UID: \"8e70e48b-fe5b-4292-b609-baec526562df\") " Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.714732 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-config" (OuterVolumeSpecName: "config") pod "92bf50f8-1f71-4cf9-9f36-7209ab28dc5e" (UID: "92bf50f8-1f71-4cf9-9f36-7209ab28dc5e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.715202 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e70e48b-fe5b-4292-b609-baec526562df-config" (OuterVolumeSpecName: "config") pod "8e70e48b-fe5b-4292-b609-baec526562df" (UID: "8e70e48b-fe5b-4292-b609-baec526562df"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.715726 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.715753 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8e70e48b-fe5b-4292-b609-baec526562df-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.716774 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" event={"ID":"92bf50f8-1f71-4cf9-9f36-7209ab28dc5e","Type":"ContainerDied","Data":"a8c9aa9547826c1e1014f4184fa4e3fa5ff6628f97f9c7f9260cb4ce4b2f32ec"} Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.716793 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-5dkxq" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.716846 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "92bf50f8-1f71-4cf9-9f36-7209ab28dc5e" (UID: "92bf50f8-1f71-4cf9-9f36-7209ab28dc5e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.718546 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" event={"ID":"8e70e48b-fe5b-4292-b609-baec526562df","Type":"ContainerDied","Data":"a65c626c011c6638c7eda1d99ea332b14e5811e4120bd36c6c07eb81695a6de8"} Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.718643 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-s8cx2" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.728010 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-kube-api-access-jpk2q" (OuterVolumeSpecName: "kube-api-access-jpk2q") pod "92bf50f8-1f71-4cf9-9f36-7209ab28dc5e" (UID: "92bf50f8-1f71-4cf9-9f36-7209ab28dc5e"). InnerVolumeSpecName "kube-api-access-jpk2q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.728120 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zw9cp" event={"ID":"20dc8601-3918-4d9c-993e-5ec14c0140b4","Type":"ContainerStarted","Data":"516bbd1690087ef0519c8f88cb6138a929accfb87fb50b48ee2f0cbdde4c0559"} Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.729350 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3459cebe-06a7-428e-93ef-419677fbcb58","Type":"ContainerStarted","Data":"cc92050992e5d80e3d17bbbee7ae4b6da86e6bd6bcc3ddabf1d1a8d5c51b5d1f"} Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.733970 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"267a25c3-20d9-4b6d-b097-d14b9534faa1","Type":"ContainerStarted","Data":"d17eb4e70e7492b1c006c38a129a9180fafabb7d947347a58c0d5d76b3b15fea"} Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.761811 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e70e48b-fe5b-4292-b609-baec526562df-kube-api-access-nxnq7" (OuterVolumeSpecName: "kube-api-access-nxnq7") pod "8e70e48b-fe5b-4292-b609-baec526562df" (UID: "8e70e48b-fe5b-4292-b609-baec526562df"). InnerVolumeSpecName "kube-api-access-nxnq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.818364 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.818694 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jpk2q\" (UniqueName: \"kubernetes.io/projected/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e-kube-api-access-jpk2q\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:05 crc kubenswrapper[4930]: I0313 09:33:05.818821 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nxnq7\" (UniqueName: \"kubernetes.io/projected/8e70e48b-fe5b-4292-b609-baec526562df-kube-api-access-nxnq7\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:06 crc kubenswrapper[4930]: I0313 09:33:06.182502 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s8cx2"] Mar 13 09:33:06 crc kubenswrapper[4930]: I0313 09:33:06.230285 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-s8cx2"] Mar 13 09:33:06 crc kubenswrapper[4930]: I0313 09:33:06.263845 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5dkxq"] Mar 13 09:33:06 crc kubenswrapper[4930]: I0313 09:33:06.275280 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-5dkxq"] Mar 13 09:33:06 crc kubenswrapper[4930]: I0313 09:33:06.283188 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 13 09:33:06 crc kubenswrapper[4930]: I0313 09:33:06.401972 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 13 09:33:06 crc kubenswrapper[4930]: I0313 09:33:06.742153 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6e3a3580-5ee8-4853-8b81-43768ea76d20","Type":"ContainerStarted","Data":"45227330525a113498f2fea0442c0f08e3874f1c94862450e0ce7b73fb075fde"} Mar 13 09:33:06 crc kubenswrapper[4930]: I0313 09:33:06.744615 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"2f69fa32-5b72-4f9b-9176-d43ee011605d","Type":"ContainerStarted","Data":"f5bb5cadafeb7d370421ec827fb7acdb9c3d0c47530810cf1b6e9e27c93891de"} Mar 13 09:33:06 crc kubenswrapper[4930]: I0313 09:33:06.750468 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7656bb8967-5vbb4" event={"ID":"91a9c833-bbb1-4341-899b-d922b4066dc0","Type":"ContainerStarted","Data":"f8d94b25364c5b095aa56613345a2b6e246aa3b132a9875afaede44395624bec"} Mar 13 09:33:06 crc kubenswrapper[4930]: I0313 09:33:06.752091 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354","Type":"ContainerStarted","Data":"43fb9ce2d8aa5390976a443491d5cf85b350d3f0f72289a47f81a4a7564022d1"} Mar 13 09:33:06 crc kubenswrapper[4930]: I0313 09:33:06.753162 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d08733c1-5776-499e-8dbe-1cc29dd9716e","Type":"ContainerStarted","Data":"3b71d8fdddeabe359d23ab2a5c71458d094cbd999398ccf75638b2e5ba6bc0e9"} Mar 13 09:33:06 crc kubenswrapper[4930]: I0313 09:33:06.831258 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-7656bb8967-5vbb4" podStartSLOduration=27.831231969 podStartE2EDuration="27.831231969s" podCreationTimestamp="2026-03-13 09:32:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:06.817531204 +0000 UTC m=+1227.567445881" watchObservedRunningTime="2026-03-13 09:33:06.831231969 +0000 UTC m=+1227.581146646" Mar 13 09:33:07 crc kubenswrapper[4930]: I0313 09:33:07.766102 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"897aa260-40c2-42d2-b59f-964a1a40116c","Type":"ContainerStarted","Data":"5c636643df2852c587885134c9b4420f836bd5a85a714cb57ef6dc4fd70b578a"} Mar 13 09:33:07 crc kubenswrapper[4930]: I0313 09:33:07.982562 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e70e48b-fe5b-4292-b609-baec526562df" path="/var/lib/kubelet/pods/8e70e48b-fe5b-4292-b609-baec526562df/volumes" Mar 13 09:33:07 crc kubenswrapper[4930]: I0313 09:33:07.982939 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92bf50f8-1f71-4cf9-9f36-7209ab28dc5e" path="/var/lib/kubelet/pods/92bf50f8-1f71-4cf9-9f36-7209ab28dc5e/volumes" Mar 13 09:33:09 crc kubenswrapper[4930]: I0313 09:33:09.782687 4930 generic.go:334] "Generic (PLEG): container finished" podID="3459cebe-06a7-428e-93ef-419677fbcb58" containerID="cc92050992e5d80e3d17bbbee7ae4b6da86e6bd6bcc3ddabf1d1a8d5c51b5d1f" exitCode=0 Mar 13 09:33:09 crc kubenswrapper[4930]: I0313 09:33:09.782779 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3459cebe-06a7-428e-93ef-419677fbcb58","Type":"ContainerDied","Data":"cc92050992e5d80e3d17bbbee7ae4b6da86e6bd6bcc3ddabf1d1a8d5c51b5d1f"} Mar 13 09:33:09 crc kubenswrapper[4930]: I0313 09:33:09.836856 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:33:09 crc kubenswrapper[4930]: I0313 09:33:09.836936 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:33:09 crc kubenswrapper[4930]: I0313 09:33:09.840841 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:33:10 crc kubenswrapper[4930]: I0313 09:33:10.795381 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 09:33:10 crc kubenswrapper[4930]: I0313 09:33:10.874095 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-7b97b57845-4zjvn"] Mar 13 09:33:11 crc kubenswrapper[4930]: I0313 09:33:11.809584 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0","Type":"ContainerStarted","Data":"9e3b5ddea5078998a0e4923edb408014e41fbcd484e03ca1e4e4372dc642a263"} Mar 13 09:33:12 crc kubenswrapper[4930]: I0313 09:33:12.826289 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3459cebe-06a7-428e-93ef-419677fbcb58","Type":"ContainerStarted","Data":"b85c79faf68f9744648c98621a97e4ae9f30c296fa782f54ee36a56172c622e1"} Mar 13 09:33:12 crc kubenswrapper[4930]: I0313 09:33:12.829898 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6e3a3580-5ee8-4853-8b81-43768ea76d20","Type":"ContainerStarted","Data":"b5197ec0659259cc9ce05c16f518231609c36cb91711dfe51d957f0828df6fae"} Mar 13 09:33:12 crc kubenswrapper[4930]: I0313 09:33:12.833932 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9b328c28-0be8-4f5c-8e76-bf33c8ee1726","Type":"ContainerStarted","Data":"f28d0b2c01d1baceaa9f6ecc2a63545f0b997feb57264f88a6518935fd0246a3"} Mar 13 09:33:12 crc kubenswrapper[4930]: I0313 09:33:12.834734 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 13 09:33:12 crc kubenswrapper[4930]: I0313 09:33:12.837907 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8" event={"ID":"5e86c9f9-c149-46a6-9210-ec642593320d","Type":"ContainerStarted","Data":"c06bdb82d6fb00e7b4ea47452b2d2775a4e69cee2fae09937dfa57e22a330e56"} Mar 13 09:33:12 crc kubenswrapper[4930]: I0313 09:33:12.842217 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ecb88f7d-f623-4395-9c32-566a3ef1f0c9","Type":"ContainerStarted","Data":"30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f"} Mar 13 09:33:12 crc kubenswrapper[4930]: I0313 09:33:12.843113 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 09:33:12 crc kubenswrapper[4930]: I0313 09:33:12.855849 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=13.173843005 podStartE2EDuration="38.855828262s" podCreationTimestamp="2026-03-13 09:32:34 +0000 UTC" firstStartedPulling="2026-03-13 09:32:36.45198112 +0000 UTC m=+1197.201895797" lastFinishedPulling="2026-03-13 09:33:02.133966377 +0000 UTC m=+1222.883881054" observedRunningTime="2026-03-13 09:33:12.846494208 +0000 UTC m=+1233.596408885" watchObservedRunningTime="2026-03-13 09:33:12.855828262 +0000 UTC m=+1233.605742939" Mar 13 09:33:12 crc kubenswrapper[4930]: I0313 09:33:12.859994 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d08733c1-5776-499e-8dbe-1cc29dd9716e","Type":"ContainerStarted","Data":"6e840d8ded438e5647867a328d7387379dafc7df9e83ffa238d5c2797fda59ab"} Mar 13 09:33:12 crc kubenswrapper[4930]: I0313 09:33:12.878690 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=10.466330332 podStartE2EDuration="34.878666327s" podCreationTimestamp="2026-03-13 09:32:38 +0000 UTC" firstStartedPulling="2026-03-13 09:32:46.900896269 +0000 UTC m=+1207.650810946" lastFinishedPulling="2026-03-13 09:33:11.313232274 +0000 UTC m=+1232.063146941" observedRunningTime="2026-03-13 09:33:12.862038519 +0000 UTC m=+1233.611953196" watchObservedRunningTime="2026-03-13 09:33:12.878666327 +0000 UTC m=+1233.628581004" Mar 13 09:33:12 crc kubenswrapper[4930]: I0313 09:33:12.891335 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.197335802 podStartE2EDuration="37.891313875s" podCreationTimestamp="2026-03-13 09:32:35 +0000 UTC" firstStartedPulling="2026-03-13 09:32:36.668797955 +0000 UTC m=+1197.418712632" lastFinishedPulling="2026-03-13 09:33:12.362776028 +0000 UTC m=+1233.112690705" observedRunningTime="2026-03-13 09:33:12.881011696 +0000 UTC m=+1233.630926383" watchObservedRunningTime="2026-03-13 09:33:12.891313875 +0000 UTC m=+1233.641228552" Mar 13 09:33:12 crc kubenswrapper[4930]: I0313 09:33:12.908502 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-66cbf594b5-kpvf8" podStartSLOduration=27.985398009 podStartE2EDuration="34.908479437s" podCreationTimestamp="2026-03-13 09:32:38 +0000 UTC" firstStartedPulling="2026-03-13 09:33:05.420047506 +0000 UTC m=+1226.169962183" lastFinishedPulling="2026-03-13 09:33:12.343128934 +0000 UTC m=+1233.093043611" observedRunningTime="2026-03-13 09:33:12.894083415 +0000 UTC m=+1233.643998122" watchObservedRunningTime="2026-03-13 09:33:12.908479437 +0000 UTC m=+1233.658394114" Mar 13 09:33:13 crc kubenswrapper[4930]: I0313 09:33:13.872737 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj" event={"ID":"2307735f-0b44-41ed-a622-74d62c05810c","Type":"ContainerStarted","Data":"d5f0bff3ca0daa7c3f0aa0f65ee7425d8fe57366dc70eb3c7f2c656c82fec80d"} Mar 13 09:33:13 crc kubenswrapper[4930]: I0313 09:33:13.873072 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-fghdj" Mar 13 09:33:13 crc kubenswrapper[4930]: I0313 09:33:13.879242 4930 generic.go:334] "Generic (PLEG): container finished" podID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerID="0fd8c22b548d89b7279f0a7b633330ae5ace09acb2da71ba19396d02612fd826" exitCode=0 Mar 13 09:33:13 crc kubenswrapper[4930]: I0313 09:33:13.880105 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zw9cp" event={"ID":"20dc8601-3918-4d9c-993e-5ec14c0140b4","Type":"ContainerDied","Data":"0fd8c22b548d89b7279f0a7b633330ae5ace09acb2da71ba19396d02612fd826"} Mar 13 09:33:13 crc kubenswrapper[4930]: I0313 09:33:13.911482 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-fghdj" podStartSLOduration=24.391330822 podStartE2EDuration="31.9114589s" podCreationTimestamp="2026-03-13 09:32:42 +0000 UTC" firstStartedPulling="2026-03-13 09:33:04.842564448 +0000 UTC m=+1225.592479115" lastFinishedPulling="2026-03-13 09:33:12.362692516 +0000 UTC m=+1233.112607193" observedRunningTime="2026-03-13 09:33:13.891229501 +0000 UTC m=+1234.641144178" watchObservedRunningTime="2026-03-13 09:33:13.9114589 +0000 UTC m=+1234.661373577" Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.803239 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-lg6gm"] Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.805069 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.810082 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.820705 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lg6gm"] Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.898631 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a6623f09-f961-411a-9323-8cbf953c7148","Type":"ContainerStarted","Data":"96c39c7b929d20787c563592a28f65306475af49da6b2538643645d233ad69d4"} Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.903216 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zw9cp" event={"ID":"20dc8601-3918-4d9c-993e-5ec14c0140b4","Type":"ContainerStarted","Data":"3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba"} Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.928026 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3281f3a8-8651-448c-abe8-70acad219523-ovn-rundir\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.928085 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95r5f\" (UniqueName: \"kubernetes.io/projected/3281f3a8-8651-448c-abe8-70acad219523-kube-api-access-95r5f\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.928107 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3281f3a8-8651-448c-abe8-70acad219523-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.928228 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3281f3a8-8651-448c-abe8-70acad219523-ovs-rundir\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.928283 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3281f3a8-8651-448c-abe8-70acad219523-config\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.928300 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3281f3a8-8651-448c-abe8-70acad219523-combined-ca-bundle\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.960767 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-zc722"] Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.992473 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-tq4bj"] Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.994600 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:14 crc kubenswrapper[4930]: I0313 09:33:14.999706 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.025921 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-tq4bj"] Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.038672 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3281f3a8-8651-448c-abe8-70acad219523-config\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.038724 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3281f3a8-8651-448c-abe8-70acad219523-combined-ca-bundle\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.038862 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3281f3a8-8651-448c-abe8-70acad219523-ovn-rundir\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.038917 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95r5f\" (UniqueName: \"kubernetes.io/projected/3281f3a8-8651-448c-abe8-70acad219523-kube-api-access-95r5f\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.038970 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3281f3a8-8651-448c-abe8-70acad219523-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.039290 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3281f3a8-8651-448c-abe8-70acad219523-ovs-rundir\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.041908 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3281f3a8-8651-448c-abe8-70acad219523-config\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.045347 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/3281f3a8-8651-448c-abe8-70acad219523-ovn-rundir\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.050449 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/3281f3a8-8651-448c-abe8-70acad219523-ovs-rundir\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.114205 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3281f3a8-8651-448c-abe8-70acad219523-combined-ca-bundle\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.119539 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95r5f\" (UniqueName: \"kubernetes.io/projected/3281f3a8-8651-448c-abe8-70acad219523-kube-api-access-95r5f\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.129066 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/3281f3a8-8651-448c-abe8-70acad219523-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-lg6gm\" (UID: \"3281f3a8-8651-448c-abe8-70acad219523\") " pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.148507 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-tq4bj\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.148635 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-config\") pod \"dnsmasq-dns-6bc7876d45-tq4bj\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.148728 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6stfm\" (UniqueName: \"kubernetes.io/projected/f74242dd-a009-415a-bd4f-91d5df4e4e08-kube-api-access-6stfm\") pod \"dnsmasq-dns-6bc7876d45-tq4bj\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.148988 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-tq4bj\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.250765 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-tq4bj\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.251228 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-tq4bj\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.251276 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-config\") pod \"dnsmasq-dns-6bc7876d45-tq4bj\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.251323 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6stfm\" (UniqueName: \"kubernetes.io/projected/f74242dd-a009-415a-bd4f-91d5df4e4e08-kube-api-access-6stfm\") pod \"dnsmasq-dns-6bc7876d45-tq4bj\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.258328 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-tq4bj\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.259004 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-config\") pod \"dnsmasq-dns-6bc7876d45-tq4bj\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.262417 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-tq4bj\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.317170 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6stfm\" (UniqueName: \"kubernetes.io/projected/f74242dd-a009-415a-bd4f-91d5df4e4e08-kube-api-access-6stfm\") pod \"dnsmasq-dns-6bc7876d45-tq4bj\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.377620 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.429835 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-lg6gm" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.441471 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zsn7s"] Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.483776 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-nkhpc"] Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.485411 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.499747 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.509476 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-nkhpc"] Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.557705 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klh98\" (UniqueName: \"kubernetes.io/projected/af4031c0-0506-40a7-a5ae-2a3bb91a5521-kube-api-access-klh98\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.558015 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.558144 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.558266 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-config\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.558536 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-dns-svc\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.561639 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.562778 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.661111 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-dns-svc\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.661192 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klh98\" (UniqueName: \"kubernetes.io/projected/af4031c0-0506-40a7-a5ae-2a3bb91a5521-kube-api-access-klh98\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.661222 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.661259 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.661298 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-config\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.662483 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-config\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.662585 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-dns-svc\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.662742 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.663007 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.684279 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klh98\" (UniqueName: \"kubernetes.io/projected/af4031c0-0506-40a7-a5ae-2a3bb91a5521-kube-api-access-klh98\") pod \"dnsmasq-dns-8554648995-nkhpc\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.857502 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.913420 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"267a25c3-20d9-4b6d-b097-d14b9534faa1","Type":"ContainerStarted","Data":"36254e9d39ddea7c93391459ef1c96007498228c864b0042412472d83f0d7b83"} Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.916580 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zw9cp" event={"ID":"20dc8601-3918-4d9c-993e-5ec14c0140b4","Type":"ContainerStarted","Data":"783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8"} Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.916629 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.916641 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:33:15 crc kubenswrapper[4930]: I0313 09:33:15.966113 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-zw9cp" podStartSLOduration=27.230129628 podStartE2EDuration="33.966039067s" podCreationTimestamp="2026-03-13 09:32:42 +0000 UTC" firstStartedPulling="2026-03-13 09:33:05.444762148 +0000 UTC m=+1226.194676825" lastFinishedPulling="2026-03-13 09:33:12.180671587 +0000 UTC m=+1232.930586264" observedRunningTime="2026-03-13 09:33:15.960190883 +0000 UTC m=+1236.710105570" watchObservedRunningTime="2026-03-13 09:33:15.966039067 +0000 UTC m=+1236.715953734" Mar 13 09:33:16 crc kubenswrapper[4930]: E0313 09:33:16.918654 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8adee7e3_81a7_4dcf_ad19_d50e1900d3e0.slice/crio-9e3b5ddea5078998a0e4923edb408014e41fbcd484e03ca1e4e4372dc642a263.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8adee7e3_81a7_4dcf_ad19_d50e1900d3e0.slice/crio-conmon-9e3b5ddea5078998a0e4923edb408014e41fbcd484e03ca1e4e4372dc642a263.scope\": RecentStats: unable to find data in memory cache]" Mar 13 09:33:16 crc kubenswrapper[4930]: I0313 09:33:16.925561 4930 generic.go:334] "Generic (PLEG): container finished" podID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerID="9e3b5ddea5078998a0e4923edb408014e41fbcd484e03ca1e4e4372dc642a263" exitCode=0 Mar 13 09:33:16 crc kubenswrapper[4930]: I0313 09:33:16.926525 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0","Type":"ContainerDied","Data":"9e3b5ddea5078998a0e4923edb408014e41fbcd484e03ca1e4e4372dc642a263"} Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.575808 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.582123 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.701556 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba559492-1275-4e57-ba68-b94af80e931b-dns-svc\") pod \"ba559492-1275-4e57-ba68-b94af80e931b\" (UID: \"ba559492-1275-4e57-ba68-b94af80e931b\") " Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.701613 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9hwlr\" (UniqueName: \"kubernetes.io/projected/ba559492-1275-4e57-ba68-b94af80e931b-kube-api-access-9hwlr\") pod \"ba559492-1275-4e57-ba68-b94af80e931b\" (UID: \"ba559492-1275-4e57-ba68-b94af80e931b\") " Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.701693 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba559492-1275-4e57-ba68-b94af80e931b-config\") pod \"ba559492-1275-4e57-ba68-b94af80e931b\" (UID: \"ba559492-1275-4e57-ba68-b94af80e931b\") " Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.701812 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qjdhf\" (UniqueName: \"kubernetes.io/projected/b302d040-3268-4633-b651-560d5a0645d2-kube-api-access-qjdhf\") pod \"b302d040-3268-4633-b651-560d5a0645d2\" (UID: \"b302d040-3268-4633-b651-560d5a0645d2\") " Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.701877 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b302d040-3268-4633-b651-560d5a0645d2-dns-svc\") pod \"b302d040-3268-4633-b651-560d5a0645d2\" (UID: \"b302d040-3268-4633-b651-560d5a0645d2\") " Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.701936 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b302d040-3268-4633-b651-560d5a0645d2-config\") pod \"b302d040-3268-4633-b651-560d5a0645d2\" (UID: \"b302d040-3268-4633-b651-560d5a0645d2\") " Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.702108 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba559492-1275-4e57-ba68-b94af80e931b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ba559492-1275-4e57-ba68-b94af80e931b" (UID: "ba559492-1275-4e57-ba68-b94af80e931b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.702258 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba559492-1275-4e57-ba68-b94af80e931b-config" (OuterVolumeSpecName: "config") pod "ba559492-1275-4e57-ba68-b94af80e931b" (UID: "ba559492-1275-4e57-ba68-b94af80e931b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.702588 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b302d040-3268-4633-b651-560d5a0645d2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b302d040-3268-4633-b651-560d5a0645d2" (UID: "b302d040-3268-4633-b651-560d5a0645d2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.702767 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b302d040-3268-4633-b651-560d5a0645d2-config" (OuterVolumeSpecName: "config") pod "b302d040-3268-4633-b651-560d5a0645d2" (UID: "b302d040-3268-4633-b651-560d5a0645d2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.703115 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba559492-1275-4e57-ba68-b94af80e931b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.703129 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba559492-1275-4e57-ba68-b94af80e931b-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.703138 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b302d040-3268-4633-b651-560d5a0645d2-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.703146 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b302d040-3268-4633-b651-560d5a0645d2-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.724514 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b302d040-3268-4633-b651-560d5a0645d2-kube-api-access-qjdhf" (OuterVolumeSpecName: "kube-api-access-qjdhf") pod "b302d040-3268-4633-b651-560d5a0645d2" (UID: "b302d040-3268-4633-b651-560d5a0645d2"). InnerVolumeSpecName "kube-api-access-qjdhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.725890 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba559492-1275-4e57-ba68-b94af80e931b-kube-api-access-9hwlr" (OuterVolumeSpecName: "kube-api-access-9hwlr") pod "ba559492-1275-4e57-ba68-b94af80e931b" (UID: "ba559492-1275-4e57-ba68-b94af80e931b"). InnerVolumeSpecName "kube-api-access-9hwlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.805151 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9hwlr\" (UniqueName: \"kubernetes.io/projected/ba559492-1275-4e57-ba68-b94af80e931b-kube-api-access-9hwlr\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.805186 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qjdhf\" (UniqueName: \"kubernetes.io/projected/b302d040-3268-4633-b651-560d5a0645d2-kube-api-access-qjdhf\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.934387 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.934394 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-zc722" event={"ID":"b302d040-3268-4633-b651-560d5a0645d2","Type":"ContainerDied","Data":"1f528838203ea0e6aa144f33f506efef885dd25f696e5faa83b463524e812027"} Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.935820 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" event={"ID":"ba559492-1275-4e57-ba68-b94af80e931b","Type":"ContainerDied","Data":"54cf96892290030dd9da19e9a5ae367c5d0b7159c943e58ab5de9b12c026fa36"} Mar 13 09:33:17 crc kubenswrapper[4930]: I0313 09:33:17.936279 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-zsn7s" Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.039998 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zsn7s"] Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.054562 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-zsn7s"] Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.078892 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-zc722"] Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.103550 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-zc722"] Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.376592 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.428864 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-nkhpc"] Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.537404 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.607162 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.784239 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-tq4bj"] Mar 13 09:33:18 crc kubenswrapper[4930]: W0313 09:33:18.793658 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf74242dd_a009_415a_bd4f_91d5df4e4e08.slice/crio-fc107dd06ffe04b52f941345497361e22b9bd825182100fbc8dcbb220900a5cc WatchSource:0}: Error finding container fc107dd06ffe04b52f941345497361e22b9bd825182100fbc8dcbb220900a5cc: Status 404 returned error can't find the container with id fc107dd06ffe04b52f941345497361e22b9bd825182100fbc8dcbb220900a5cc Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.904235 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-lg6gm"] Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.944738 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0","Type":"ContainerStarted","Data":"c3c777a41680fafe3cd112f765e67426a68e7274d8368d2cff35e95608324d32"} Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.948299 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" event={"ID":"f74242dd-a009-415a-bd4f-91d5df4e4e08","Type":"ContainerStarted","Data":"fc107dd06ffe04b52f941345497361e22b9bd825182100fbc8dcbb220900a5cc"} Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.950212 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lg6gm" event={"ID":"3281f3a8-8651-448c-abe8-70acad219523","Type":"ContainerStarted","Data":"9add9e7f7d87eba6798111f82fd282b4b30e83cdb445a1e849b7df95caad23ca"} Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.952391 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"d08733c1-5776-499e-8dbe-1cc29dd9716e","Type":"ContainerStarted","Data":"1ed1a84d035eeecd641d78eb98424c04638ac738161e000ef21fb32a528ae22a"} Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.957511 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6e3a3580-5ee8-4853-8b81-43768ea76d20","Type":"ContainerStarted","Data":"f8bdc118e946c25eaf16fd412967cb18a276dadf4e1b4d2aac8764ec937143e6"} Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.958832 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-nkhpc" event={"ID":"af4031c0-0506-40a7-a5ae-2a3bb91a5521","Type":"ContainerStarted","Data":"9aa9e95d5a87075e73986c4fd0d098911715e9c0b467f2e42cbc7427d7b0bc53"} Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.975391 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=17.8024358 podStartE2EDuration="46.97536711s" podCreationTimestamp="2026-03-13 09:32:32 +0000 UTC" firstStartedPulling="2026-03-13 09:32:35.132822203 +0000 UTC m=+1195.882736890" lastFinishedPulling="2026-03-13 09:33:04.305753533 +0000 UTC m=+1225.055668200" observedRunningTime="2026-03-13 09:33:18.962979756 +0000 UTC m=+1239.712894433" watchObservedRunningTime="2026-03-13 09:33:18.97536711 +0000 UTC m=+1239.725281807" Mar 13 09:33:18 crc kubenswrapper[4930]: I0313 09:33:18.998757 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=27.158482177 podStartE2EDuration="38.998739152s" podCreationTimestamp="2026-03-13 09:32:40 +0000 UTC" firstStartedPulling="2026-03-13 09:33:06.584239895 +0000 UTC m=+1227.334154582" lastFinishedPulling="2026-03-13 09:33:18.42449688 +0000 UTC m=+1239.174411557" observedRunningTime="2026-03-13 09:33:18.987636341 +0000 UTC m=+1239.737551018" watchObservedRunningTime="2026-03-13 09:33:18.998739152 +0000 UTC m=+1239.748653829" Mar 13 09:33:19 crc kubenswrapper[4930]: I0313 09:33:19.011804 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=24.186813607 podStartE2EDuration="36.011789954s" podCreationTimestamp="2026-03-13 09:32:43 +0000 UTC" firstStartedPulling="2026-03-13 09:33:06.585795854 +0000 UTC m=+1227.335710531" lastFinishedPulling="2026-03-13 09:33:18.410772201 +0000 UTC m=+1239.160686878" observedRunningTime="2026-03-13 09:33:19.011507236 +0000 UTC m=+1239.761421923" watchObservedRunningTime="2026-03-13 09:33:19.011789954 +0000 UTC m=+1239.761704631" Mar 13 09:33:19 crc kubenswrapper[4930]: I0313 09:33:19.967734 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-lg6gm" event={"ID":"3281f3a8-8651-448c-abe8-70acad219523","Type":"ContainerStarted","Data":"b843e614f6ad6759ab09f5cf6104fa7056901dc686f5266c5d86ab8fd138d61c"} Mar 13 09:33:19 crc kubenswrapper[4930]: I0313 09:33:19.972217 4930 generic.go:334] "Generic (PLEG): container finished" podID="af4031c0-0506-40a7-a5ae-2a3bb91a5521" containerID="3275277c91bfd35d1dafe4301b570a319af9112e60a19cde36448f4e939fba29" exitCode=0 Mar 13 09:33:19 crc kubenswrapper[4930]: I0313 09:33:19.974941 4930 generic.go:334] "Generic (PLEG): container finished" podID="f74242dd-a009-415a-bd4f-91d5df4e4e08" containerID="d721efb0df8ac0a5a237a7ee81278bbdc30ddcd9047eb8d44140edc2aa41c47f" exitCode=0 Mar 13 09:33:19 crc kubenswrapper[4930]: I0313 09:33:19.988294 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b302d040-3268-4633-b651-560d5a0645d2" path="/var/lib/kubelet/pods/b302d040-3268-4633-b651-560d5a0645d2/volumes" Mar 13 09:33:19 crc kubenswrapper[4930]: I0313 09:33:19.988785 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba559492-1275-4e57-ba68-b94af80e931b" path="/var/lib/kubelet/pods/ba559492-1275-4e57-ba68-b94af80e931b/volumes" Mar 13 09:33:19 crc kubenswrapper[4930]: I0313 09:33:19.989288 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 13 09:33:19 crc kubenswrapper[4930]: I0313 09:33:19.989311 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-nkhpc" event={"ID":"af4031c0-0506-40a7-a5ae-2a3bb91a5521","Type":"ContainerDied","Data":"3275277c91bfd35d1dafe4301b570a319af9112e60a19cde36448f4e939fba29"} Mar 13 09:33:19 crc kubenswrapper[4930]: I0313 09:33:19.989328 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" event={"ID":"f74242dd-a009-415a-bd4f-91d5df4e4e08","Type":"ContainerDied","Data":"d721efb0df8ac0a5a237a7ee81278bbdc30ddcd9047eb8d44140edc2aa41c47f"} Mar 13 09:33:19 crc kubenswrapper[4930]: I0313 09:33:19.992321 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-lg6gm" podStartSLOduration=5.99230044 podStartE2EDuration="5.99230044s" podCreationTimestamp="2026-03-13 09:33:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:19.986008066 +0000 UTC m=+1240.735922743" watchObservedRunningTime="2026-03-13 09:33:19.99230044 +0000 UTC m=+1240.742215117" Mar 13 09:33:20 crc kubenswrapper[4930]: I0313 09:33:20.845333 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 13 09:33:20 crc kubenswrapper[4930]: I0313 09:33:20.892251 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 13 09:33:20 crc kubenswrapper[4930]: I0313 09:33:20.923623 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 13 09:33:20 crc kubenswrapper[4930]: I0313 09:33:20.974703 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 13 09:33:21 crc kubenswrapper[4930]: I0313 09:33:21.028899 4930 generic.go:334] "Generic (PLEG): container finished" podID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerID="36254e9d39ddea7c93391459ef1c96007498228c864b0042412472d83f0d7b83" exitCode=0 Mar 13 09:33:21 crc kubenswrapper[4930]: I0313 09:33:21.029053 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"267a25c3-20d9-4b6d-b097-d14b9534faa1","Type":"ContainerDied","Data":"36254e9d39ddea7c93391459ef1c96007498228c864b0042412472d83f0d7b83"} Mar 13 09:33:21 crc kubenswrapper[4930]: I0313 09:33:21.044647 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-nkhpc" event={"ID":"af4031c0-0506-40a7-a5ae-2a3bb91a5521","Type":"ContainerStarted","Data":"e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8"} Mar 13 09:33:21 crc kubenswrapper[4930]: I0313 09:33:21.044729 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:21 crc kubenswrapper[4930]: I0313 09:33:21.050644 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" event={"ID":"f74242dd-a009-415a-bd4f-91d5df4e4e08","Type":"ContainerStarted","Data":"13740b3d775b84f1d7ac0f12767ef57fb9b1f8d4de1cd51da0134da50bbdb1fa"} Mar 13 09:33:21 crc kubenswrapper[4930]: I0313 09:33:21.051189 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 13 09:33:21 crc kubenswrapper[4930]: I0313 09:33:21.051724 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 13 09:33:21 crc kubenswrapper[4930]: I0313 09:33:21.051819 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:21 crc kubenswrapper[4930]: I0313 09:33:21.110756 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" podStartSLOduration=6.54583745 podStartE2EDuration="7.110734207s" podCreationTimestamp="2026-03-13 09:33:14 +0000 UTC" firstStartedPulling="2026-03-13 09:33:18.796484918 +0000 UTC m=+1239.546399585" lastFinishedPulling="2026-03-13 09:33:19.361381665 +0000 UTC m=+1240.111296342" observedRunningTime="2026-03-13 09:33:21.11007404 +0000 UTC m=+1241.859988727" watchObservedRunningTime="2026-03-13 09:33:21.110734207 +0000 UTC m=+1241.860648894" Mar 13 09:33:21 crc kubenswrapper[4930]: I0313 09:33:21.114485 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 13 09:33:21 crc kubenswrapper[4930]: I0313 09:33:21.142081 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-nkhpc" podStartSLOduration=5.614024125 podStartE2EDuration="6.142061627s" podCreationTimestamp="2026-03-13 09:33:15 +0000 UTC" firstStartedPulling="2026-03-13 09:33:18.449918896 +0000 UTC m=+1239.199833573" lastFinishedPulling="2026-03-13 09:33:18.977956398 +0000 UTC m=+1239.727871075" observedRunningTime="2026-03-13 09:33:21.141228276 +0000 UTC m=+1241.891142953" watchObservedRunningTime="2026-03-13 09:33:21.142061627 +0000 UTC m=+1241.891976304" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.105013 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.426028 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.427885 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.430829 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-nl456" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.431011 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.431199 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.431325 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.458163 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.547017 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c09afa02-80df-434d-999c-58e6787d0d0d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.547101 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c09afa02-80df-434d-999c-58e6787d0d0d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.547142 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c09afa02-80df-434d-999c-58e6787d0d0d-scripts\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.547254 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c09afa02-80df-434d-999c-58e6787d0d0d-config\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.547321 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c09afa02-80df-434d-999c-58e6787d0d0d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.547361 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09afa02-80df-434d-999c-58e6787d0d0d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.547383 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5tlfk\" (UniqueName: \"kubernetes.io/projected/c09afa02-80df-434d-999c-58e6787d0d0d-kube-api-access-5tlfk\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.648720 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c09afa02-80df-434d-999c-58e6787d0d0d-config\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.648819 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c09afa02-80df-434d-999c-58e6787d0d0d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.648855 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09afa02-80df-434d-999c-58e6787d0d0d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.648879 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5tlfk\" (UniqueName: \"kubernetes.io/projected/c09afa02-80df-434d-999c-58e6787d0d0d-kube-api-access-5tlfk\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.648942 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c09afa02-80df-434d-999c-58e6787d0d0d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.648993 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c09afa02-80df-434d-999c-58e6787d0d0d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.649028 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c09afa02-80df-434d-999c-58e6787d0d0d-scripts\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.649540 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c09afa02-80df-434d-999c-58e6787d0d0d-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.649969 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c09afa02-80df-434d-999c-58e6787d0d0d-scripts\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.650317 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c09afa02-80df-434d-999c-58e6787d0d0d-config\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.655819 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c09afa02-80df-434d-999c-58e6787d0d0d-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.666691 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c09afa02-80df-434d-999c-58e6787d0d0d-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.668712 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c09afa02-80df-434d-999c-58e6787d0d0d-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.669850 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5tlfk\" (UniqueName: \"kubernetes.io/projected/c09afa02-80df-434d-999c-58e6787d0d0d-kube-api-access-5tlfk\") pod \"ovn-northd-0\" (UID: \"c09afa02-80df-434d-999c-58e6787d0d0d\") " pod="openstack/ovn-northd-0" Mar 13 09:33:22 crc kubenswrapper[4930]: I0313 09:33:22.768570 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 13 09:33:23 crc kubenswrapper[4930]: I0313 09:33:23.250609 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 13 09:33:23 crc kubenswrapper[4930]: W0313 09:33:23.264574 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc09afa02_80df_434d_999c_58e6787d0d0d.slice/crio-8bede9c9bcde189d88c98239be11699a057d6d445ddc4a3095cc25bb25240e1c WatchSource:0}: Error finding container 8bede9c9bcde189d88c98239be11699a057d6d445ddc4a3095cc25bb25240e1c: Status 404 returned error can't find the container with id 8bede9c9bcde189d88c98239be11699a057d6d445ddc4a3095cc25bb25240e1c Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.092313 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c09afa02-80df-434d-999c-58e6787d0d0d","Type":"ContainerStarted","Data":"8bede9c9bcde189d88c98239be11699a057d6d445ddc4a3095cc25bb25240e1c"} Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.249057 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-qnkx9"] Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.250660 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-qnkx9" Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.262813 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.265022 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-qnkx9"] Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.295608 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.295744 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.395811 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ff7901f-6245-437b-a7de-aa3e338fac88-operator-scripts\") pod \"root-account-create-update-qnkx9\" (UID: \"5ff7901f-6245-437b-a7de-aa3e338fac88\") " pod="openstack/root-account-create-update-qnkx9" Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.395899 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mt4n8\" (UniqueName: \"kubernetes.io/projected/5ff7901f-6245-437b-a7de-aa3e338fac88-kube-api-access-mt4n8\") pod \"root-account-create-update-qnkx9\" (UID: \"5ff7901f-6245-437b-a7de-aa3e338fac88\") " pod="openstack/root-account-create-update-qnkx9" Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.497736 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ff7901f-6245-437b-a7de-aa3e338fac88-operator-scripts\") pod \"root-account-create-update-qnkx9\" (UID: \"5ff7901f-6245-437b-a7de-aa3e338fac88\") " pod="openstack/root-account-create-update-qnkx9" Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.497822 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mt4n8\" (UniqueName: \"kubernetes.io/projected/5ff7901f-6245-437b-a7de-aa3e338fac88-kube-api-access-mt4n8\") pod \"root-account-create-update-qnkx9\" (UID: \"5ff7901f-6245-437b-a7de-aa3e338fac88\") " pod="openstack/root-account-create-update-qnkx9" Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.498929 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ff7901f-6245-437b-a7de-aa3e338fac88-operator-scripts\") pod \"root-account-create-update-qnkx9\" (UID: \"5ff7901f-6245-437b-a7de-aa3e338fac88\") " pod="openstack/root-account-create-update-qnkx9" Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.529314 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mt4n8\" (UniqueName: \"kubernetes.io/projected/5ff7901f-6245-437b-a7de-aa3e338fac88-kube-api-access-mt4n8\") pod \"root-account-create-update-qnkx9\" (UID: \"5ff7901f-6245-437b-a7de-aa3e338fac88\") " pod="openstack/root-account-create-update-qnkx9" Mar 13 09:33:24 crc kubenswrapper[4930]: I0313 09:33:24.583899 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-qnkx9" Mar 13 09:33:25 crc kubenswrapper[4930]: I0313 09:33:25.200053 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-qnkx9"] Mar 13 09:33:25 crc kubenswrapper[4930]: I0313 09:33:25.380251 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:25 crc kubenswrapper[4930]: I0313 09:33:25.859728 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:25 crc kubenswrapper[4930]: I0313 09:33:25.958625 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-tq4bj"] Mar 13 09:33:26 crc kubenswrapper[4930]: I0313 09:33:26.115309 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-qnkx9" event={"ID":"5ff7901f-6245-437b-a7de-aa3e338fac88","Type":"ContainerStarted","Data":"522a5aa45c21c167b50a4a7d68c2e4d86489c97a3e3bdc9d915d6bef9e626fec"} Mar 13 09:33:26 crc kubenswrapper[4930]: I0313 09:33:26.115464 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" podUID="f74242dd-a009-415a-bd4f-91d5df4e4e08" containerName="dnsmasq-dns" containerID="cri-o://13740b3d775b84f1d7ac0f12767ef57fb9b1f8d4de1cd51da0134da50bbdb1fa" gracePeriod=10 Mar 13 09:33:27 crc kubenswrapper[4930]: I0313 09:33:27.123871 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-qnkx9" event={"ID":"5ff7901f-6245-437b-a7de-aa3e338fac88","Type":"ContainerStarted","Data":"12f5210598b95db7dbe92711d04a787ad629536cd28bb916dd8442d27c127c93"} Mar 13 09:33:27 crc kubenswrapper[4930]: I0313 09:33:27.126607 4930 generic.go:334] "Generic (PLEG): container finished" podID="f74242dd-a009-415a-bd4f-91d5df4e4e08" containerID="13740b3d775b84f1d7ac0f12767ef57fb9b1f8d4de1cd51da0134da50bbdb1fa" exitCode=0 Mar 13 09:33:27 crc kubenswrapper[4930]: I0313 09:33:27.126649 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" event={"ID":"f74242dd-a009-415a-bd4f-91d5df4e4e08","Type":"ContainerDied","Data":"13740b3d775b84f1d7ac0f12767ef57fb9b1f8d4de1cd51da0134da50bbdb1fa"} Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.336059 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-g4vq8"] Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.339153 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.357345 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-g4vq8"] Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.388514 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.388610 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frb69\" (UniqueName: \"kubernetes.io/projected/f4d00778-de97-4b50-a752-42109b188b11-kube-api-access-frb69\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.388719 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-config\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.388751 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.388772 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.490756 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.490884 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frb69\" (UniqueName: \"kubernetes.io/projected/f4d00778-de97-4b50-a752-42109b188b11-kube-api-access-frb69\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.491017 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-config\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.491054 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.491073 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.491738 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.491933 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.492026 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.492097 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-config\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.520968 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frb69\" (UniqueName: \"kubernetes.io/projected/f4d00778-de97-4b50-a752-42109b188b11-kube-api-access-frb69\") pod \"dnsmasq-dns-b8fbc5445-g4vq8\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.663364 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:28 crc kubenswrapper[4930]: I0313 09:33:28.959963 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.052748 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerName="galera" probeResult="failure" output=< Mar 13 09:33:29 crc kubenswrapper[4930]: wsrep_local_state_comment (Joined) differs from Synced Mar 13 09:33:29 crc kubenswrapper[4930]: > Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.119346 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-g4vq8"] Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.146360 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" event={"ID":"f4d00778-de97-4b50-a752-42109b188b11","Type":"ContainerStarted","Data":"73b864a8044621e56716f7bf6195d4e3ebe32abbad9ded54e4bed89fc9fab815"} Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.175800 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-qnkx9" podStartSLOduration=5.175783002 podStartE2EDuration="5.175783002s" podCreationTimestamp="2026-03-13 09:33:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:29.167601747 +0000 UTC m=+1249.917516424" watchObservedRunningTime="2026-03-13 09:33:29.175783002 +0000 UTC m=+1249.925697679" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.499577 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.506019 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.515550 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.515580 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.515694 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-ljrpm" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.515874 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.516094 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.617251 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b288113-babd-4a27-b397-cbdcbfdc195c-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.617544 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.617599 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7b288113-babd-4a27-b397-cbdcbfdc195c-cache\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.617649 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-931333d3-ad4d-499f-b198-49c28d4c78fd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-931333d3-ad4d-499f-b198-49c28d4c78fd\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.617685 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7b288113-babd-4a27-b397-cbdcbfdc195c-lock\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.617743 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b7p4k\" (UniqueName: \"kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-kube-api-access-b7p4k\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.720056 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b7p4k\" (UniqueName: \"kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-kube-api-access-b7p4k\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.720565 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b288113-babd-4a27-b397-cbdcbfdc195c-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.720694 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: E0313 09:33:29.720834 4930 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 09:33:29 crc kubenswrapper[4930]: E0313 09:33:29.720876 4930 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 09:33:29 crc kubenswrapper[4930]: E0313 09:33:29.720947 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift podName:7b288113-babd-4a27-b397-cbdcbfdc195c nodeName:}" failed. No retries permitted until 2026-03-13 09:33:30.220920452 +0000 UTC m=+1250.970835129 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift") pod "swift-storage-0" (UID: "7b288113-babd-4a27-b397-cbdcbfdc195c") : configmap "swift-ring-files" not found Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.721055 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7b288113-babd-4a27-b397-cbdcbfdc195c-cache\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.721225 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-931333d3-ad4d-499f-b198-49c28d4c78fd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-931333d3-ad4d-499f-b198-49c28d4c78fd\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.721411 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7b288113-babd-4a27-b397-cbdcbfdc195c-lock\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.721603 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/7b288113-babd-4a27-b397-cbdcbfdc195c-cache\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.721973 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/7b288113-babd-4a27-b397-cbdcbfdc195c-lock\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.726639 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b288113-babd-4a27-b397-cbdcbfdc195c-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.769270 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b7p4k\" (UniqueName: \"kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-kube-api-access-b7p4k\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.926231 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.926277 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-931333d3-ad4d-499f-b198-49c28d4c78fd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-931333d3-ad4d-499f-b198-49c28d4c78fd\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/db2a99ff71a12b2ad9168ec491f7108b035885fcc8022e611f79b36800866998/globalmount\"" pod="openstack/swift-storage-0" Mar 13 09:33:29 crc kubenswrapper[4930]: I0313 09:33:29.962774 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-931333d3-ad4d-499f-b198-49c28d4c78fd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-931333d3-ad4d-499f-b198-49c28d4c78fd\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.142805 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-nqsgg"] Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.144894 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.148263 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.148713 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.148903 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.153704 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-nqsgg"] Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.182800 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4d00778-de97-4b50-a752-42109b188b11" containerID="52428e5ba5283d2d616708463dc9171cdbca2df3692703b0a5d6fc5413f59f49" exitCode=0 Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.182850 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" event={"ID":"f4d00778-de97-4b50-a752-42109b188b11","Type":"ContainerDied","Data":"52428e5ba5283d2d616708463dc9171cdbca2df3692703b0a5d6fc5413f59f49"} Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.242928 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-dispersionconf\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.243128 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-combined-ca-bundle\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.243381 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-swiftconf\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.243454 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2230cfae-6fbf-4df4-a249-622fe3c957e3-etc-swift\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.243669 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2230cfae-6fbf-4df4-a249-622fe3c957e3-scripts\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.243710 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbqgl\" (UniqueName: \"kubernetes.io/projected/2230cfae-6fbf-4df4-a249-622fe3c957e3-kube-api-access-xbqgl\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.243826 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2230cfae-6fbf-4df4-a249-622fe3c957e3-ring-data-devices\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.243897 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:30 crc kubenswrapper[4930]: E0313 09:33:30.244099 4930 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 09:33:30 crc kubenswrapper[4930]: E0313 09:33:30.244112 4930 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 09:33:30 crc kubenswrapper[4930]: E0313 09:33:30.244149 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift podName:7b288113-babd-4a27-b397-cbdcbfdc195c nodeName:}" failed. No retries permitted until 2026-03-13 09:33:31.244135988 +0000 UTC m=+1251.994050665 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift") pod "swift-storage-0" (UID: "7b288113-babd-4a27-b397-cbdcbfdc195c") : configmap "swift-ring-files" not found Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.345569 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2230cfae-6fbf-4df4-a249-622fe3c957e3-scripts\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.345629 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbqgl\" (UniqueName: \"kubernetes.io/projected/2230cfae-6fbf-4df4-a249-622fe3c957e3-kube-api-access-xbqgl\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.345693 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2230cfae-6fbf-4df4-a249-622fe3c957e3-ring-data-devices\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.345749 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-dispersionconf\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.345790 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-combined-ca-bundle\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.345927 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-swiftconf\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.345959 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2230cfae-6fbf-4df4-a249-622fe3c957e3-etc-swift\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.346331 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2230cfae-6fbf-4df4-a249-622fe3c957e3-etc-swift\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.346425 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2230cfae-6fbf-4df4-a249-622fe3c957e3-scripts\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.346585 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2230cfae-6fbf-4df4-a249-622fe3c957e3-ring-data-devices\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.352019 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-dispersionconf\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.352584 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-combined-ca-bundle\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.365983 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-swiftconf\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.369557 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbqgl\" (UniqueName: \"kubernetes.io/projected/2230cfae-6fbf-4df4-a249-622fe3c957e3-kube-api-access-xbqgl\") pod \"swift-ring-rebalance-nqsgg\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:30 crc kubenswrapper[4930]: I0313 09:33:30.481985 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:31 crc kubenswrapper[4930]: I0313 09:33:31.209394 4930 generic.go:334] "Generic (PLEG): container finished" podID="5ff7901f-6245-437b-a7de-aa3e338fac88" containerID="12f5210598b95db7dbe92711d04a787ad629536cd28bb916dd8442d27c127c93" exitCode=0 Mar 13 09:33:31 crc kubenswrapper[4930]: I0313 09:33:31.209476 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-qnkx9" event={"ID":"5ff7901f-6245-437b-a7de-aa3e338fac88","Type":"ContainerDied","Data":"12f5210598b95db7dbe92711d04a787ad629536cd28bb916dd8442d27c127c93"} Mar 13 09:33:31 crc kubenswrapper[4930]: I0313 09:33:31.269013 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:31 crc kubenswrapper[4930]: E0313 09:33:31.269246 4930 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 09:33:31 crc kubenswrapper[4930]: E0313 09:33:31.269273 4930 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 09:33:31 crc kubenswrapper[4930]: E0313 09:33:31.269340 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift podName:7b288113-babd-4a27-b397-cbdcbfdc195c nodeName:}" failed. No retries permitted until 2026-03-13 09:33:33.269317714 +0000 UTC m=+1254.019232391 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift") pod "swift-storage-0" (UID: "7b288113-babd-4a27-b397-cbdcbfdc195c") : configmap "swift-ring-files" not found Mar 13 09:33:33 crc kubenswrapper[4930]: I0313 09:33:33.313348 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:33 crc kubenswrapper[4930]: E0313 09:33:33.313695 4930 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 09:33:33 crc kubenswrapper[4930]: E0313 09:33:33.313933 4930 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 09:33:33 crc kubenswrapper[4930]: E0313 09:33:33.313999 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift podName:7b288113-babd-4a27-b397-cbdcbfdc195c nodeName:}" failed. No retries permitted until 2026-03-13 09:33:37.313979095 +0000 UTC m=+1258.063893772 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift") pod "swift-storage-0" (UID: "7b288113-babd-4a27-b397-cbdcbfdc195c") : configmap "swift-ring-files" not found Mar 13 09:33:34 crc kubenswrapper[4930]: I0313 09:33:34.378249 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 13 09:33:35 crc kubenswrapper[4930]: I0313 09:33:35.380288 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" podUID="f74242dd-a009-415a-bd4f-91d5df4e4e08" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.150:5353: i/o timeout" Mar 13 09:33:35 crc kubenswrapper[4930]: I0313 09:33:35.974165 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-7b97b57845-4zjvn" podUID="7c5b562e-e1d3-4a2d-8169-e4b8a687251a" containerName="console" containerID="cri-o://36e5fa63122da9c19f41e3bf274a7c37b2bb9f6a39f72e7e6bff6f02d133cb71" gracePeriod=15 Mar 13 09:33:35 crc kubenswrapper[4930]: E0313 09:33:35.992097 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:1b555e21bba7c609111ace4380382a696d9aceeb6e9816bf9023b8f689b6c741" Mar 13 09:33:35 crc kubenswrapper[4930]: E0313 09:33:35.992325 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:prometheus,Image:registry.redhat.io/cluster-observability-operator/prometheus-rhel9@sha256:1b555e21bba7c609111ace4380382a696d9aceeb6e9816bf9023b8f689b6c741,Command:[],Args:[--config.file=/etc/prometheus/config_out/prometheus.env.yaml --web.enable-lifecycle --web.route-prefix=/ --storage.tsdb.retention.time=24h --storage.tsdb.path=/prometheus --web.config.file=/etc/prometheus/web_config/web-config.yaml],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:web,HostPort:0,ContainerPort:9090,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-out,ReadOnly:true,MountPath:/etc/prometheus/config_out,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tls-assets,ReadOnly:true,MountPath:/etc/prometheus/certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:prometheus-metric-storage-db,ReadOnly:false,MountPath:/prometheus,SubPath:prometheus-db,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:prometheus-metric-storage-rulefiles-0,ReadOnly:true,MountPath:/etc/prometheus/rules/prometheus-metric-storage-rulefiles-0,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:prometheus-metric-storage-rulefiles-1,ReadOnly:true,MountPath:/etc/prometheus/rules/prometheus-metric-storage-rulefiles-1,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:prometheus-metric-storage-rulefiles-2,ReadOnly:true,MountPath:/etc/prometheus/rules/prometheus-metric-storage-rulefiles-2,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:web-config,ReadOnly:true,MountPath:/etc/prometheus/web_config/web-config.yaml,SubPath:web-config.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9mvlv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/-/healthy,Port:{1 0 web},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:3,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:6,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/-/ready,Port:{1 0 web},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:3,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*true,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/-/ready,Port:{1 0 web},Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:3,PeriodSeconds:15,SuccessThreshold:1,FailureThreshold:60,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod prometheus-metric-storage-0_openstack(267a25c3-20d9-4b6d-b097-d14b9534faa1): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.250280 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7b97b57845-4zjvn_7c5b562e-e1d3-4a2d-8169-e4b8a687251a/console/0.log" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.250325 4930 generic.go:334] "Generic (PLEG): container finished" podID="7c5b562e-e1d3-4a2d-8169-e4b8a687251a" containerID="36e5fa63122da9c19f41e3bf274a7c37b2bb9f6a39f72e7e6bff6f02d133cb71" exitCode=2 Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.250352 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7b97b57845-4zjvn" event={"ID":"7c5b562e-e1d3-4a2d-8169-e4b8a687251a","Type":"ContainerDied","Data":"36e5fa63122da9c19f41e3bf274a7c37b2bb9f6a39f72e7e6bff6f02d133cb71"} Mar 13 09:33:36 crc kubenswrapper[4930]: E0313 09:33:36.275061 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified" Mar 13 09:33:36 crc kubenswrapper[4930]: E0313 09:33:36.275644 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ovn-northd,Image:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,Command:[/usr/bin/ovn-northd],Args:[-vfile:off -vconsole:info --n-threads=1 --ovnnb-db=ssl:ovsdbserver-nb-0.openstack.svc.cluster.local:6641 --ovnsb-db=ssl:ovsdbserver-sb-0.openstack.svc.cluster.local:6642 --certificate=/etc/pki/tls/certs/ovndb.crt --private-key=/etc/pki/tls/private/ovndb.key --ca-cert=/etc/pki/tls/certs/ovndbca.crt],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nch76h574h5f5h679h97h8dh69h4hd6h98hfch5b4h59chc5hd6hcfh5f9hd6hb4h544h5fchd7h5dch6bh655h68dh664h6h549h576h6dq,ValueFrom:nil,},EnvVar{Name:OVN_RUNDIR,Value:/tmp,ValueFrom:nil,},EnvVar{Name:certs,Value:n86h7fh57bh688h4h694h656h5b9hc9h676h7ch565h5fhd8h9ch644h584h67ch7bh5c8h549hcch4h5bbhbch546h695h97h648h5f5h9ch665q,ValueFrom:nil,},EnvVar{Name:certs_metrics,Value:n84hb6hffh85h84h659h5b5hc7hd9h5dbh58bh577h559hc5h576h66bh547h645h86h676h55ch554h556h5bch664h5ch64dh667h668h5ch76h8cq,ValueFrom:nil,},EnvVar{Name:ovnnorthd-config,Value:n5c8h7ch56bh8dh8hc4h5dch9dh68h6bhb7h598h549h5dbh66fh6bh5b4h5cch5d6h55ch57fhfch588h89h5ddh5d6h65bh65bh8dhc4h67dh569q,ValueFrom:nil,},EnvVar{Name:ovnnorthd-scripts,Value:n664hd8h66ch58dh64hc9h66bhd4h558h697h67bh557hdch664h567h669h555h696h556h556h5fh5bh569hbh665h9dh4h9bh564hc8h5b7h5c4q,ValueFrom:nil,},EnvVar{Name:tls-ca-bundle.pem,Value:n676h545h55fh65fh5dh579h88h569h6dh67fh9fh58fh5fch5dch645h55bh9bhd6h585h9bhbhcfh6dh68fh95h67bh78h557h576hdfh7hfbq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-rundir,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-northd-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndb.crt,SubPath:tls.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-northd-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/private/ovndb.key,SubPath:tls.key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovn-northd-tls-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ovndbca.crt,SubPath:ca.crt,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5tlfk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/status_check.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:15,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/local/bin/container-scripts/status_check.sh],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:15,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ovn-northd-0_openstack(c09afa02-80df-434d-999c-58e6787d0d0d): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.496740 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.530316 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-qnkx9" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.623876 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mt4n8\" (UniqueName: \"kubernetes.io/projected/5ff7901f-6245-437b-a7de-aa3e338fac88-kube-api-access-mt4n8\") pod \"5ff7901f-6245-437b-a7de-aa3e338fac88\" (UID: \"5ff7901f-6245-437b-a7de-aa3e338fac88\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.623935 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6stfm\" (UniqueName: \"kubernetes.io/projected/f74242dd-a009-415a-bd4f-91d5df4e4e08-kube-api-access-6stfm\") pod \"f74242dd-a009-415a-bd4f-91d5df4e4e08\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.624096 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-dns-svc\") pod \"f74242dd-a009-415a-bd4f-91d5df4e4e08\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.624158 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-ovsdbserver-sb\") pod \"f74242dd-a009-415a-bd4f-91d5df4e4e08\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.624223 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ff7901f-6245-437b-a7de-aa3e338fac88-operator-scripts\") pod \"5ff7901f-6245-437b-a7de-aa3e338fac88\" (UID: \"5ff7901f-6245-437b-a7de-aa3e338fac88\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.624252 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-config\") pod \"f74242dd-a009-415a-bd4f-91d5df4e4e08\" (UID: \"f74242dd-a009-415a-bd4f-91d5df4e4e08\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.626846 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ff7901f-6245-437b-a7de-aa3e338fac88-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5ff7901f-6245-437b-a7de-aa3e338fac88" (UID: "5ff7901f-6245-437b-a7de-aa3e338fac88"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.665588 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ff7901f-6245-437b-a7de-aa3e338fac88-kube-api-access-mt4n8" (OuterVolumeSpecName: "kube-api-access-mt4n8") pod "5ff7901f-6245-437b-a7de-aa3e338fac88" (UID: "5ff7901f-6245-437b-a7de-aa3e338fac88"). InnerVolumeSpecName "kube-api-access-mt4n8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.670151 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f74242dd-a009-415a-bd4f-91d5df4e4e08-kube-api-access-6stfm" (OuterVolumeSpecName: "kube-api-access-6stfm") pod "f74242dd-a009-415a-bd4f-91d5df4e4e08" (UID: "f74242dd-a009-415a-bd4f-91d5df4e4e08"). InnerVolumeSpecName "kube-api-access-6stfm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: E0313 09:33:36.711810 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-northd\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ovn-northd-0" podUID="c09afa02-80df-434d-999c-58e6787d0d0d" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.727493 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mt4n8\" (UniqueName: \"kubernetes.io/projected/5ff7901f-6245-437b-a7de-aa3e338fac88-kube-api-access-mt4n8\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.727712 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6stfm\" (UniqueName: \"kubernetes.io/projected/f74242dd-a009-415a-bd4f-91d5df4e4e08-kube-api-access-6stfm\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.727793 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ff7901f-6245-437b-a7de-aa3e338fac88-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.737700 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-config" (OuterVolumeSpecName: "config") pod "f74242dd-a009-415a-bd4f-91d5df4e4e08" (UID: "f74242dd-a009-415a-bd4f-91d5df4e4e08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.749698 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f74242dd-a009-415a-bd4f-91d5df4e4e08" (UID: "f74242dd-a009-415a-bd4f-91d5df4e4e08"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.777090 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f74242dd-a009-415a-bd4f-91d5df4e4e08" (UID: "f74242dd-a009-415a-bd4f-91d5df4e4e08"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.791831 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7b97b57845-4zjvn_7c5b562e-e1d3-4a2d-8169-e4b8a687251a/console/0.log" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.791915 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.830113 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.830144 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.830155 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f74242dd-a009-415a-bd4f-91d5df4e4e08-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.929121 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-k6ktf"] Mar 13 09:33:36 crc kubenswrapper[4930]: E0313 09:33:36.929558 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f74242dd-a009-415a-bd4f-91d5df4e4e08" containerName="init" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.929573 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f74242dd-a009-415a-bd4f-91d5df4e4e08" containerName="init" Mar 13 09:33:36 crc kubenswrapper[4930]: E0313 09:33:36.929597 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ff7901f-6245-437b-a7de-aa3e338fac88" containerName="mariadb-account-create-update" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.929606 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ff7901f-6245-437b-a7de-aa3e338fac88" containerName="mariadb-account-create-update" Mar 13 09:33:36 crc kubenswrapper[4930]: E0313 09:33:36.929625 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f74242dd-a009-415a-bd4f-91d5df4e4e08" containerName="dnsmasq-dns" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.929634 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f74242dd-a009-415a-bd4f-91d5df4e4e08" containerName="dnsmasq-dns" Mar 13 09:33:36 crc kubenswrapper[4930]: E0313 09:33:36.929663 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c5b562e-e1d3-4a2d-8169-e4b8a687251a" containerName="console" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.929675 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c5b562e-e1d3-4a2d-8169-e4b8a687251a" containerName="console" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.929858 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f74242dd-a009-415a-bd4f-91d5df4e4e08" containerName="dnsmasq-dns" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.929870 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ff7901f-6245-437b-a7de-aa3e338fac88" containerName="mariadb-account-create-update" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.929887 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c5b562e-e1d3-4a2d-8169-e4b8a687251a" containerName="console" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.930575 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-k6ktf" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.930814 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-trusted-ca-bundle\") pod \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.930861 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-config\") pod \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.930884 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-oauth-config\") pod \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.930952 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-service-ca\") pod \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.931044 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-oauth-serving-cert\") pod \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.931143 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5tb7\" (UniqueName: \"kubernetes.io/projected/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-kube-api-access-s5tb7\") pod \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.931169 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-serving-cert\") pod \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\" (UID: \"7c5b562e-e1d3-4a2d-8169-e4b8a687251a\") " Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.931660 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "7c5b562e-e1d3-4a2d-8169-e4b8a687251a" (UID: "7c5b562e-e1d3-4a2d-8169-e4b8a687251a"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.931682 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-config" (OuterVolumeSpecName: "console-config") pod "7c5b562e-e1d3-4a2d-8169-e4b8a687251a" (UID: "7c5b562e-e1d3-4a2d-8169-e4b8a687251a"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.931670 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-service-ca" (OuterVolumeSpecName: "service-ca") pod "7c5b562e-e1d3-4a2d-8169-e4b8a687251a" (UID: "7c5b562e-e1d3-4a2d-8169-e4b8a687251a"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.931707 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "7c5b562e-e1d3-4a2d-8169-e4b8a687251a" (UID: "7c5b562e-e1d3-4a2d-8169-e4b8a687251a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.935027 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "7c5b562e-e1d3-4a2d-8169-e4b8a687251a" (UID: "7c5b562e-e1d3-4a2d-8169-e4b8a687251a"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.935050 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-kube-api-access-s5tb7" (OuterVolumeSpecName: "kube-api-access-s5tb7") pod "7c5b562e-e1d3-4a2d-8169-e4b8a687251a" (UID: "7c5b562e-e1d3-4a2d-8169-e4b8a687251a"). InnerVolumeSpecName "kube-api-access-s5tb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.937932 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "7c5b562e-e1d3-4a2d-8169-e4b8a687251a" (UID: "7c5b562e-e1d3-4a2d-8169-e4b8a687251a"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.942080 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-nqsgg"] Mar 13 09:33:36 crc kubenswrapper[4930]: I0313 09:33:36.979686 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-k6ktf"] Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.034742 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc952175-d111-4e8c-972d-e26cca11ba49-operator-scripts\") pod \"keystone-db-create-k6ktf\" (UID: \"cc952175-d111-4e8c-972d-e26cca11ba49\") " pod="openstack/keystone-db-create-k6ktf" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.034987 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2m29t\" (UniqueName: \"kubernetes.io/projected/cc952175-d111-4e8c-972d-e26cca11ba49-kube-api-access-2m29t\") pod \"keystone-db-create-k6ktf\" (UID: \"cc952175-d111-4e8c-972d-e26cca11ba49\") " pod="openstack/keystone-db-create-k6ktf" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.035083 4930 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.035101 4930 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.035111 4930 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.035134 4930 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-service-ca\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.035143 4930 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.035153 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5tb7\" (UniqueName: \"kubernetes.io/projected/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-kube-api-access-s5tb7\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.035164 4930 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/7c5b562e-e1d3-4a2d-8169-e4b8a687251a-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.038759 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-0a6a-account-create-update-kvkqr"] Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.040188 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0a6a-account-create-update-kvkqr" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.042332 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.050033 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-0a6a-account-create-update-kvkqr"] Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.137753 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eebc86bf-e7ee-4465-bd15-bef833ac0843-operator-scripts\") pod \"keystone-0a6a-account-create-update-kvkqr\" (UID: \"eebc86bf-e7ee-4465-bd15-bef833ac0843\") " pod="openstack/keystone-0a6a-account-create-update-kvkqr" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.137902 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2m29t\" (UniqueName: \"kubernetes.io/projected/cc952175-d111-4e8c-972d-e26cca11ba49-kube-api-access-2m29t\") pod \"keystone-db-create-k6ktf\" (UID: \"cc952175-d111-4e8c-972d-e26cca11ba49\") " pod="openstack/keystone-db-create-k6ktf" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.137937 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzqdz\" (UniqueName: \"kubernetes.io/projected/eebc86bf-e7ee-4465-bd15-bef833ac0843-kube-api-access-fzqdz\") pod \"keystone-0a6a-account-create-update-kvkqr\" (UID: \"eebc86bf-e7ee-4465-bd15-bef833ac0843\") " pod="openstack/keystone-0a6a-account-create-update-kvkqr" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.138024 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc952175-d111-4e8c-972d-e26cca11ba49-operator-scripts\") pod \"keystone-db-create-k6ktf\" (UID: \"cc952175-d111-4e8c-972d-e26cca11ba49\") " pod="openstack/keystone-db-create-k6ktf" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.138956 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc952175-d111-4e8c-972d-e26cca11ba49-operator-scripts\") pod \"keystone-db-create-k6ktf\" (UID: \"cc952175-d111-4e8c-972d-e26cca11ba49\") " pod="openstack/keystone-db-create-k6ktf" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.156769 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2m29t\" (UniqueName: \"kubernetes.io/projected/cc952175-d111-4e8c-972d-e26cca11ba49-kube-api-access-2m29t\") pod \"keystone-db-create-k6ktf\" (UID: \"cc952175-d111-4e8c-972d-e26cca11ba49\") " pod="openstack/keystone-db-create-k6ktf" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.175556 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-hvlz5"] Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.177186 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hvlz5" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.197476 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-hvlz5"] Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.239652 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzqdz\" (UniqueName: \"kubernetes.io/projected/eebc86bf-e7ee-4465-bd15-bef833ac0843-kube-api-access-fzqdz\") pod \"keystone-0a6a-account-create-update-kvkqr\" (UID: \"eebc86bf-e7ee-4465-bd15-bef833ac0843\") " pod="openstack/keystone-0a6a-account-create-update-kvkqr" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.239716 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v2l5d\" (UniqueName: \"kubernetes.io/projected/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48-kube-api-access-v2l5d\") pod \"placement-db-create-hvlz5\" (UID: \"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48\") " pod="openstack/placement-db-create-hvlz5" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.239751 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48-operator-scripts\") pod \"placement-db-create-hvlz5\" (UID: \"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48\") " pod="openstack/placement-db-create-hvlz5" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.240200 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eebc86bf-e7ee-4465-bd15-bef833ac0843-operator-scripts\") pod \"keystone-0a6a-account-create-update-kvkqr\" (UID: \"eebc86bf-e7ee-4465-bd15-bef833ac0843\") " pod="openstack/keystone-0a6a-account-create-update-kvkqr" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.241149 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eebc86bf-e7ee-4465-bd15-bef833ac0843-operator-scripts\") pod \"keystone-0a6a-account-create-update-kvkqr\" (UID: \"eebc86bf-e7ee-4465-bd15-bef833ac0843\") " pod="openstack/keystone-0a6a-account-create-update-kvkqr" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.246920 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5f25-account-create-update-t9v5n"] Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.248317 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5f25-account-create-update-t9v5n" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.250740 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.268109 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzqdz\" (UniqueName: \"kubernetes.io/projected/eebc86bf-e7ee-4465-bd15-bef833ac0843-kube-api-access-fzqdz\") pod \"keystone-0a6a-account-create-update-kvkqr\" (UID: \"eebc86bf-e7ee-4465-bd15-bef833ac0843\") " pod="openstack/keystone-0a6a-account-create-update-kvkqr" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.270862 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nqsgg" event={"ID":"2230cfae-6fbf-4df4-a249-622fe3c957e3","Type":"ContainerStarted","Data":"416aef5b6b5eeda5dd38a6d87a26acba20e41cb21b52f85ecde82d776429f6ad"} Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.270875 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-k6ktf" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.273389 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c09afa02-80df-434d-999c-58e6787d0d0d","Type":"ContainerStarted","Data":"6cfa008ff507fd91ed2946c4d2ffe0c77511e03343fee3f372c336f2eeef0210"} Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.280000 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5f25-account-create-update-t9v5n"] Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.281095 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" event={"ID":"f4d00778-de97-4b50-a752-42109b188b11","Type":"ContainerStarted","Data":"9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7"} Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.281241 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.282563 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-qnkx9" event={"ID":"5ff7901f-6245-437b-a7de-aa3e338fac88","Type":"ContainerDied","Data":"522a5aa45c21c167b50a4a7d68c2e4d86489c97a3e3bdc9d915d6bef9e626fec"} Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.282590 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="522a5aa45c21c167b50a4a7d68c2e4d86489c97a3e3bdc9d915d6bef9e626fec" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.282664 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-qnkx9" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.284406 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7b97b57845-4zjvn_7c5b562e-e1d3-4a2d-8169-e4b8a687251a/console/0.log" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.284511 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7b97b57845-4zjvn" event={"ID":"7c5b562e-e1d3-4a2d-8169-e4b8a687251a","Type":"ContainerDied","Data":"3fc82d44e00d139493a20dc6e1f6848aa1eaa78e449a44127ea3e5759d6edddc"} Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.284535 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-7b97b57845-4zjvn" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.284546 4930 scope.go:117] "RemoveContainer" containerID="36e5fa63122da9c19f41e3bf274a7c37b2bb9f6a39f72e7e6bff6f02d133cb71" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.291862 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" event={"ID":"f74242dd-a009-415a-bd4f-91d5df4e4e08","Type":"ContainerDied","Data":"fc107dd06ffe04b52f941345497361e22b9bd825182100fbc8dcbb220900a5cc"} Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.291950 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" Mar 13 09:33:37 crc kubenswrapper[4930]: E0313 09:33:37.291999 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-northd\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified\\\"\"" pod="openstack/ovn-northd-0" podUID="c09afa02-80df-434d-999c-58e6787d0d0d" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.332089 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" podStartSLOduration=9.332068285 podStartE2EDuration="9.332068285s" podCreationTimestamp="2026-03-13 09:33:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:37.324790284 +0000 UTC m=+1258.074704961" watchObservedRunningTime="2026-03-13 09:33:37.332068285 +0000 UTC m=+1258.081982962" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.343060 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v2l5d\" (UniqueName: \"kubernetes.io/projected/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48-kube-api-access-v2l5d\") pod \"placement-db-create-hvlz5\" (UID: \"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48\") " pod="openstack/placement-db-create-hvlz5" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.343117 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48-operator-scripts\") pod \"placement-db-create-hvlz5\" (UID: \"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48\") " pod="openstack/placement-db-create-hvlz5" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.343251 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.343314 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-786xh\" (UniqueName: \"kubernetes.io/projected/c1731159-7f86-4cbe-b9cb-53e28cc38092-kube-api-access-786xh\") pod \"placement-5f25-account-create-update-t9v5n\" (UID: \"c1731159-7f86-4cbe-b9cb-53e28cc38092\") " pod="openstack/placement-5f25-account-create-update-t9v5n" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.343370 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1731159-7f86-4cbe-b9cb-53e28cc38092-operator-scripts\") pod \"placement-5f25-account-create-update-t9v5n\" (UID: \"c1731159-7f86-4cbe-b9cb-53e28cc38092\") " pod="openstack/placement-5f25-account-create-update-t9v5n" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.344658 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48-operator-scripts\") pod \"placement-db-create-hvlz5\" (UID: \"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48\") " pod="openstack/placement-db-create-hvlz5" Mar 13 09:33:37 crc kubenswrapper[4930]: E0313 09:33:37.345912 4930 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 09:33:37 crc kubenswrapper[4930]: E0313 09:33:37.345934 4930 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 09:33:37 crc kubenswrapper[4930]: E0313 09:33:37.345966 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift podName:7b288113-babd-4a27-b397-cbdcbfdc195c nodeName:}" failed. No retries permitted until 2026-03-13 09:33:45.345955069 +0000 UTC m=+1266.095869746 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift") pod "swift-storage-0" (UID: "7b288113-babd-4a27-b397-cbdcbfdc195c") : configmap "swift-ring-files" not found Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.364894 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v2l5d\" (UniqueName: \"kubernetes.io/projected/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48-kube-api-access-v2l5d\") pod \"placement-db-create-hvlz5\" (UID: \"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48\") " pod="openstack/placement-db-create-hvlz5" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.371008 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0a6a-account-create-update-kvkqr" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.396883 4930 scope.go:117] "RemoveContainer" containerID="13740b3d775b84f1d7ac0f12767ef57fb9b1f8d4de1cd51da0134da50bbdb1fa" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.442109 4930 scope.go:117] "RemoveContainer" containerID="d721efb0df8ac0a5a237a7ee81278bbdc30ddcd9047eb8d44140edc2aa41c47f" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.445074 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1731159-7f86-4cbe-b9cb-53e28cc38092-operator-scripts\") pod \"placement-5f25-account-create-update-t9v5n\" (UID: \"c1731159-7f86-4cbe-b9cb-53e28cc38092\") " pod="openstack/placement-5f25-account-create-update-t9v5n" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.445872 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-786xh\" (UniqueName: \"kubernetes.io/projected/c1731159-7f86-4cbe-b9cb-53e28cc38092-kube-api-access-786xh\") pod \"placement-5f25-account-create-update-t9v5n\" (UID: \"c1731159-7f86-4cbe-b9cb-53e28cc38092\") " pod="openstack/placement-5f25-account-create-update-t9v5n" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.446225 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1731159-7f86-4cbe-b9cb-53e28cc38092-operator-scripts\") pod \"placement-5f25-account-create-update-t9v5n\" (UID: \"c1731159-7f86-4cbe-b9cb-53e28cc38092\") " pod="openstack/placement-5f25-account-create-update-t9v5n" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.463570 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-786xh\" (UniqueName: \"kubernetes.io/projected/c1731159-7f86-4cbe-b9cb-53e28cc38092-kube-api-access-786xh\") pod \"placement-5f25-account-create-update-t9v5n\" (UID: \"c1731159-7f86-4cbe-b9cb-53e28cc38092\") " pod="openstack/placement-5f25-account-create-update-t9v5n" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.494071 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hvlz5" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.536640 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5f25-account-create-update-t9v5n" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.567766 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-7b97b57845-4zjvn"] Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.595411 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-7b97b57845-4zjvn"] Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.616323 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-tq4bj"] Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.643539 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-tq4bj"] Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.855760 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-k6ktf"] Mar 13 09:33:37 crc kubenswrapper[4930]: W0313 09:33:37.870718 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcc952175_d111_4e8c_972d_e26cca11ba49.slice/crio-a4c94c4c83a134d1b09c5c1b41a4c1192348380f631f567df3d8c7274c9e957a WatchSource:0}: Error finding container a4c94c4c83a134d1b09c5c1b41a4c1192348380f631f567df3d8c7274c9e957a: Status 404 returned error can't find the container with id a4c94c4c83a134d1b09c5c1b41a4c1192348380f631f567df3d8c7274c9e957a Mar 13 09:33:37 crc kubenswrapper[4930]: W0313 09:33:37.985801 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeebc86bf_e7ee_4465_bd15_bef833ac0843.slice/crio-a7b1a5751d7790154c64c4cba56830a10c7daa694f70e28037f4bed7bdabfc93 WatchSource:0}: Error finding container a7b1a5751d7790154c64c4cba56830a10c7daa694f70e28037f4bed7bdabfc93: Status 404 returned error can't find the container with id a7b1a5751d7790154c64c4cba56830a10c7daa694f70e28037f4bed7bdabfc93 Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.990028 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7c5b562e-e1d3-4a2d-8169-e4b8a687251a" path="/var/lib/kubelet/pods/7c5b562e-e1d3-4a2d-8169-e4b8a687251a/volumes" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.990846 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f74242dd-a009-415a-bd4f-91d5df4e4e08" path="/var/lib/kubelet/pods/f74242dd-a009-415a-bd4f-91d5df4e4e08/volumes" Mar 13 09:33:37 crc kubenswrapper[4930]: I0313 09:33:37.991512 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-0a6a-account-create-update-kvkqr"] Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.109141 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-hvlz5"] Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.234798 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5f25-account-create-update-t9v5n"] Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.305994 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hvlz5" event={"ID":"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48","Type":"ContainerStarted","Data":"bc1fcdcf7b5d483a3c83ea270f2062ea95f1735fff6cfe07fc663a1dd1a95bb9"} Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.307526 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-k6ktf" event={"ID":"cc952175-d111-4e8c-972d-e26cca11ba49","Type":"ContainerStarted","Data":"8f1a17ded6e885d9cba84dfb6f06c4cce0223ed0c64ea7df78c1c8c99a8e3611"} Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.307551 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-k6ktf" event={"ID":"cc952175-d111-4e8c-972d-e26cca11ba49","Type":"ContainerStarted","Data":"a4c94c4c83a134d1b09c5c1b41a4c1192348380f631f567df3d8c7274c9e957a"} Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.311196 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0a6a-account-create-update-kvkqr" event={"ID":"eebc86bf-e7ee-4465-bd15-bef833ac0843","Type":"ContainerStarted","Data":"c0c8404cd93a43566c8682cea787be016fc13cfa6d6f53223169bbcd76373fb2"} Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.311240 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0a6a-account-create-update-kvkqr" event={"ID":"eebc86bf-e7ee-4465-bd15-bef833ac0843","Type":"ContainerStarted","Data":"a7b1a5751d7790154c64c4cba56830a10c7daa694f70e28037f4bed7bdabfc93"} Mar 13 09:33:38 crc kubenswrapper[4930]: E0313 09:33:38.315380 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovn-northd\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified\\\"\"" pod="openstack/ovn-northd-0" podUID="c09afa02-80df-434d-999c-58e6787d0d0d" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.339689 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-create-k6ktf" podStartSLOduration=2.339664601 podStartE2EDuration="2.339664601s" podCreationTimestamp="2026-03-13 09:33:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:38.325296285 +0000 UTC m=+1259.075210972" watchObservedRunningTime="2026-03-13 09:33:38.339664601 +0000 UTC m=+1259.089579278" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.401175 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-0a6a-account-create-update-kvkqr" podStartSLOduration=1.40115599 podStartE2EDuration="1.40115599s" podCreationTimestamp="2026-03-13 09:33:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:38.375584351 +0000 UTC m=+1259.125499028" watchObservedRunningTime="2026-03-13 09:33:38.40115599 +0000 UTC m=+1259.151070667" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.421562 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-44qz9"] Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.423623 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-44qz9" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.440741 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-44qz9"] Mar 13 09:33:38 crc kubenswrapper[4930]: W0313 09:33:38.464025 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1731159_7f86_4cbe_b9cb_53e28cc38092.slice/crio-5faf5915d32cee9be13c21b40165a86a817e59614d60cde4c00194198925cc1f WatchSource:0}: Error finding container 5faf5915d32cee9be13c21b40165a86a817e59614d60cde4c00194198925cc1f: Status 404 returned error can't find the container with id 5faf5915d32cee9be13c21b40165a86a817e59614d60cde4c00194198925cc1f Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.571829 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/451992e0-ca47-4af8-8b22-54e3c955b3aa-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-44qz9\" (UID: \"451992e0-ca47-4af8-8b22-54e3c955b3aa\") " pod="openstack/mysqld-exporter-openstack-db-create-44qz9" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.571919 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r6fdz\" (UniqueName: \"kubernetes.io/projected/451992e0-ca47-4af8-8b22-54e3c955b3aa-kube-api-access-r6fdz\") pod \"mysqld-exporter-openstack-db-create-44qz9\" (UID: \"451992e0-ca47-4af8-8b22-54e3c955b3aa\") " pod="openstack/mysqld-exporter-openstack-db-create-44qz9" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.591708 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-5fae-account-create-update-tdfsn"] Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.593004 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.594770 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.612321 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-5fae-account-create-update-tdfsn"] Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.673763 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/451992e0-ca47-4af8-8b22-54e3c955b3aa-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-44qz9\" (UID: \"451992e0-ca47-4af8-8b22-54e3c955b3aa\") " pod="openstack/mysqld-exporter-openstack-db-create-44qz9" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.673849 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r6fdz\" (UniqueName: \"kubernetes.io/projected/451992e0-ca47-4af8-8b22-54e3c955b3aa-kube-api-access-r6fdz\") pod \"mysqld-exporter-openstack-db-create-44qz9\" (UID: \"451992e0-ca47-4af8-8b22-54e3c955b3aa\") " pod="openstack/mysqld-exporter-openstack-db-create-44qz9" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.673893 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5n4v\" (UniqueName: \"kubernetes.io/projected/b0fe8888-5e4f-4bec-a5bf-554c209325ed-kube-api-access-w5n4v\") pod \"mysqld-exporter-5fae-account-create-update-tdfsn\" (UID: \"b0fe8888-5e4f-4bec-a5bf-554c209325ed\") " pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.674016 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0fe8888-5e4f-4bec-a5bf-554c209325ed-operator-scripts\") pod \"mysqld-exporter-5fae-account-create-update-tdfsn\" (UID: \"b0fe8888-5e4f-4bec-a5bf-554c209325ed\") " pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.674655 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/451992e0-ca47-4af8-8b22-54e3c955b3aa-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-44qz9\" (UID: \"451992e0-ca47-4af8-8b22-54e3c955b3aa\") " pod="openstack/mysqld-exporter-openstack-db-create-44qz9" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.694285 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r6fdz\" (UniqueName: \"kubernetes.io/projected/451992e0-ca47-4af8-8b22-54e3c955b3aa-kube-api-access-r6fdz\") pod \"mysqld-exporter-openstack-db-create-44qz9\" (UID: \"451992e0-ca47-4af8-8b22-54e3c955b3aa\") " pod="openstack/mysqld-exporter-openstack-db-create-44qz9" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.776151 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5n4v\" (UniqueName: \"kubernetes.io/projected/b0fe8888-5e4f-4bec-a5bf-554c209325ed-kube-api-access-w5n4v\") pod \"mysqld-exporter-5fae-account-create-update-tdfsn\" (UID: \"b0fe8888-5e4f-4bec-a5bf-554c209325ed\") " pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.776358 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0fe8888-5e4f-4bec-a5bf-554c209325ed-operator-scripts\") pod \"mysqld-exporter-5fae-account-create-update-tdfsn\" (UID: \"b0fe8888-5e4f-4bec-a5bf-554c209325ed\") " pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.778172 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0fe8888-5e4f-4bec-a5bf-554c209325ed-operator-scripts\") pod \"mysqld-exporter-5fae-account-create-update-tdfsn\" (UID: \"b0fe8888-5e4f-4bec-a5bf-554c209325ed\") " pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" Mar 13 09:33:38 crc kubenswrapper[4930]: I0313 09:33:38.792277 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5n4v\" (UniqueName: \"kubernetes.io/projected/b0fe8888-5e4f-4bec-a5bf-554c209325ed-kube-api-access-w5n4v\") pod \"mysqld-exporter-5fae-account-create-update-tdfsn\" (UID: \"b0fe8888-5e4f-4bec-a5bf-554c209325ed\") " pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.026508 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-44qz9" Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.036778 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.329603 4930 generic.go:334] "Generic (PLEG): container finished" podID="dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" containerID="43fb9ce2d8aa5390976a443491d5cf85b350d3f0f72289a47f81a4a7564022d1" exitCode=0 Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.329806 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354","Type":"ContainerDied","Data":"43fb9ce2d8aa5390976a443491d5cf85b350d3f0f72289a47f81a4a7564022d1"} Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.334684 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5f25-account-create-update-t9v5n" event={"ID":"c1731159-7f86-4cbe-b9cb-53e28cc38092","Type":"ContainerStarted","Data":"5faf5915d32cee9be13c21b40165a86a817e59614d60cde4c00194198925cc1f"} Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.336406 4930 generic.go:334] "Generic (PLEG): container finished" podID="897aa260-40c2-42d2-b59f-964a1a40116c" containerID="5c636643df2852c587885134c9b4420f836bd5a85a714cb57ef6dc4fd70b578a" exitCode=0 Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.336456 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"897aa260-40c2-42d2-b59f-964a1a40116c","Type":"ContainerDied","Data":"5c636643df2852c587885134c9b4420f836bd5a85a714cb57ef6dc4fd70b578a"} Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.342349 4930 generic.go:334] "Generic (PLEG): container finished" podID="eebc86bf-e7ee-4465-bd15-bef833ac0843" containerID="c0c8404cd93a43566c8682cea787be016fc13cfa6d6f53223169bbcd76373fb2" exitCode=0 Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.342532 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0a6a-account-create-update-kvkqr" event={"ID":"eebc86bf-e7ee-4465-bd15-bef833ac0843","Type":"ContainerDied","Data":"c0c8404cd93a43566c8682cea787be016fc13cfa6d6f53223169bbcd76373fb2"} Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.352870 4930 generic.go:334] "Generic (PLEG): container finished" podID="cc952175-d111-4e8c-972d-e26cca11ba49" containerID="8f1a17ded6e885d9cba84dfb6f06c4cce0223ed0c64ea7df78c1c8c99a8e3611" exitCode=0 Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.352975 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-k6ktf" event={"ID":"cc952175-d111-4e8c-972d-e26cca11ba49","Type":"ContainerDied","Data":"8f1a17ded6e885d9cba84dfb6f06c4cce0223ed0c64ea7df78c1c8c99a8e3611"} Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.356186 4930 generic.go:334] "Generic (PLEG): container finished" podID="2f69fa32-5b72-4f9b-9176-d43ee011605d" containerID="f5bb5cadafeb7d370421ec827fb7acdb9c3d0c47530810cf1b6e9e27c93891de" exitCode=0 Mar 13 09:33:39 crc kubenswrapper[4930]: I0313 09:33:39.356239 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"2f69fa32-5b72-4f9b-9176-d43ee011605d","Type":"ContainerDied","Data":"f5bb5cadafeb7d370421ec827fb7acdb9c3d0c47530810cf1b6e9e27c93891de"} Mar 13 09:33:40 crc kubenswrapper[4930]: I0313 09:33:40.368910 4930 generic.go:334] "Generic (PLEG): container finished" podID="c1731159-7f86-4cbe-b9cb-53e28cc38092" containerID="e37e8c0afc0bbe36080c34a4600320ec76139a7878514ab22d546557609d3afc" exitCode=0 Mar 13 09:33:40 crc kubenswrapper[4930]: I0313 09:33:40.369032 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5f25-account-create-update-t9v5n" event={"ID":"c1731159-7f86-4cbe-b9cb-53e28cc38092","Type":"ContainerDied","Data":"e37e8c0afc0bbe36080c34a4600320ec76139a7878514ab22d546557609d3afc"} Mar 13 09:33:40 crc kubenswrapper[4930]: I0313 09:33:40.371835 4930 generic.go:334] "Generic (PLEG): container finished" podID="ed73f0c1-2715-409f-9a4a-9da4fa8a3f48" containerID="44c06115d18097c2cb7c6cfe31705ac6acb7f5865b83b24e0aeddcb7ffa980cf" exitCode=0 Mar 13 09:33:40 crc kubenswrapper[4930]: I0313 09:33:40.372026 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hvlz5" event={"ID":"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48","Type":"ContainerDied","Data":"44c06115d18097c2cb7c6cfe31705ac6acb7f5865b83b24e0aeddcb7ffa980cf"} Mar 13 09:33:40 crc kubenswrapper[4930]: I0313 09:33:40.384151 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6bc7876d45-tq4bj" podUID="f74242dd-a009-415a-bd4f-91d5df4e4e08" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.150:5353: i/o timeout" Mar 13 09:33:40 crc kubenswrapper[4930]: I0313 09:33:40.387252 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"267a25c3-20d9-4b6d-b097-d14b9534faa1","Type":"ContainerStarted","Data":"9aaa51fb4611379e8c17b3648f7eeeccd410a16f268c69209f0374f1300e0f58"} Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.084632 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-bqwkt"] Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.086353 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bqwkt" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.101424 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bqwkt"] Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.123738 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ssg79\" (UniqueName: \"kubernetes.io/projected/0d7d0626-3583-4461-957d-d978d8f78ed4-kube-api-access-ssg79\") pod \"glance-db-create-bqwkt\" (UID: \"0d7d0626-3583-4461-957d-d978d8f78ed4\") " pod="openstack/glance-db-create-bqwkt" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.123821 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d7d0626-3583-4461-957d-d978d8f78ed4-operator-scripts\") pod \"glance-db-create-bqwkt\" (UID: \"0d7d0626-3583-4461-957d-d978d8f78ed4\") " pod="openstack/glance-db-create-bqwkt" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.191273 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-8baa-account-create-update-b5w9f"] Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.192484 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8baa-account-create-update-b5w9f" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.200111 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.208267 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-8baa-account-create-update-b5w9f"] Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.226021 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ssg79\" (UniqueName: \"kubernetes.io/projected/0d7d0626-3583-4461-957d-d978d8f78ed4-kube-api-access-ssg79\") pod \"glance-db-create-bqwkt\" (UID: \"0d7d0626-3583-4461-957d-d978d8f78ed4\") " pod="openstack/glance-db-create-bqwkt" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.226128 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d7d0626-3583-4461-957d-d978d8f78ed4-operator-scripts\") pod \"glance-db-create-bqwkt\" (UID: \"0d7d0626-3583-4461-957d-d978d8f78ed4\") " pod="openstack/glance-db-create-bqwkt" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.227069 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d7d0626-3583-4461-957d-d978d8f78ed4-operator-scripts\") pod \"glance-db-create-bqwkt\" (UID: \"0d7d0626-3583-4461-957d-d978d8f78ed4\") " pod="openstack/glance-db-create-bqwkt" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.249580 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ssg79\" (UniqueName: \"kubernetes.io/projected/0d7d0626-3583-4461-957d-d978d8f78ed4-kube-api-access-ssg79\") pod \"glance-db-create-bqwkt\" (UID: \"0d7d0626-3583-4461-957d-d978d8f78ed4\") " pod="openstack/glance-db-create-bqwkt" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.327958 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3efe40c-96ae-4b9e-a787-7781db5af486-operator-scripts\") pod \"glance-8baa-account-create-update-b5w9f\" (UID: \"f3efe40c-96ae-4b9e-a787-7781db5af486\") " pod="openstack/glance-8baa-account-create-update-b5w9f" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.328413 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cxr5\" (UniqueName: \"kubernetes.io/projected/f3efe40c-96ae-4b9e-a787-7781db5af486-kube-api-access-2cxr5\") pod \"glance-8baa-account-create-update-b5w9f\" (UID: \"f3efe40c-96ae-4b9e-a787-7781db5af486\") " pod="openstack/glance-8baa-account-create-update-b5w9f" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.407970 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bqwkt" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.429804 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cxr5\" (UniqueName: \"kubernetes.io/projected/f3efe40c-96ae-4b9e-a787-7781db5af486-kube-api-access-2cxr5\") pod \"glance-8baa-account-create-update-b5w9f\" (UID: \"f3efe40c-96ae-4b9e-a787-7781db5af486\") " pod="openstack/glance-8baa-account-create-update-b5w9f" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.429896 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3efe40c-96ae-4b9e-a787-7781db5af486-operator-scripts\") pod \"glance-8baa-account-create-update-b5w9f\" (UID: \"f3efe40c-96ae-4b9e-a787-7781db5af486\") " pod="openstack/glance-8baa-account-create-update-b5w9f" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.430690 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3efe40c-96ae-4b9e-a787-7781db5af486-operator-scripts\") pod \"glance-8baa-account-create-update-b5w9f\" (UID: \"f3efe40c-96ae-4b9e-a787-7781db5af486\") " pod="openstack/glance-8baa-account-create-update-b5w9f" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.449479 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cxr5\" (UniqueName: \"kubernetes.io/projected/f3efe40c-96ae-4b9e-a787-7781db5af486-kube-api-access-2cxr5\") pod \"glance-8baa-account-create-update-b5w9f\" (UID: \"f3efe40c-96ae-4b9e-a787-7781db5af486\") " pod="openstack/glance-8baa-account-create-update-b5w9f" Mar 13 09:33:41 crc kubenswrapper[4930]: I0313 09:33:41.508958 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8baa-account-create-update-b5w9f" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.193684 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hvlz5" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.210355 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0a6a-account-create-update-kvkqr" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.246764 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v2l5d\" (UniqueName: \"kubernetes.io/projected/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48-kube-api-access-v2l5d\") pod \"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48\" (UID: \"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48\") " Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.246895 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48-operator-scripts\") pod \"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48\" (UID: \"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48\") " Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.247034 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzqdz\" (UniqueName: \"kubernetes.io/projected/eebc86bf-e7ee-4465-bd15-bef833ac0843-kube-api-access-fzqdz\") pod \"eebc86bf-e7ee-4465-bd15-bef833ac0843\" (UID: \"eebc86bf-e7ee-4465-bd15-bef833ac0843\") " Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.247183 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eebc86bf-e7ee-4465-bd15-bef833ac0843-operator-scripts\") pod \"eebc86bf-e7ee-4465-bd15-bef833ac0843\" (UID: \"eebc86bf-e7ee-4465-bd15-bef833ac0843\") " Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.248074 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed73f0c1-2715-409f-9a4a-9da4fa8a3f48" (UID: "ed73f0c1-2715-409f-9a4a-9da4fa8a3f48"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.249206 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eebc86bf-e7ee-4465-bd15-bef833ac0843-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "eebc86bf-e7ee-4465-bd15-bef833ac0843" (UID: "eebc86bf-e7ee-4465-bd15-bef833ac0843"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.252725 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48-kube-api-access-v2l5d" (OuterVolumeSpecName: "kube-api-access-v2l5d") pod "ed73f0c1-2715-409f-9a4a-9da4fa8a3f48" (UID: "ed73f0c1-2715-409f-9a4a-9da4fa8a3f48"). InnerVolumeSpecName "kube-api-access-v2l5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.256914 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eebc86bf-e7ee-4465-bd15-bef833ac0843-kube-api-access-fzqdz" (OuterVolumeSpecName: "kube-api-access-fzqdz") pod "eebc86bf-e7ee-4465-bd15-bef833ac0843" (UID: "eebc86bf-e7ee-4465-bd15-bef833ac0843"). InnerVolumeSpecName "kube-api-access-fzqdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.335785 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-k6ktf" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.349924 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.349952 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzqdz\" (UniqueName: \"kubernetes.io/projected/eebc86bf-e7ee-4465-bd15-bef833ac0843-kube-api-access-fzqdz\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.349963 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/eebc86bf-e7ee-4465-bd15-bef833ac0843-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.349972 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v2l5d\" (UniqueName: \"kubernetes.io/projected/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48-kube-api-access-v2l5d\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.363631 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5f25-account-create-update-t9v5n" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.451846 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-786xh\" (UniqueName: \"kubernetes.io/projected/c1731159-7f86-4cbe-b9cb-53e28cc38092-kube-api-access-786xh\") pod \"c1731159-7f86-4cbe-b9cb-53e28cc38092\" (UID: \"c1731159-7f86-4cbe-b9cb-53e28cc38092\") " Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.451937 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc952175-d111-4e8c-972d-e26cca11ba49-operator-scripts\") pod \"cc952175-d111-4e8c-972d-e26cca11ba49\" (UID: \"cc952175-d111-4e8c-972d-e26cca11ba49\") " Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.451954 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2m29t\" (UniqueName: \"kubernetes.io/projected/cc952175-d111-4e8c-972d-e26cca11ba49-kube-api-access-2m29t\") pod \"cc952175-d111-4e8c-972d-e26cca11ba49\" (UID: \"cc952175-d111-4e8c-972d-e26cca11ba49\") " Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.452854 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1731159-7f86-4cbe-b9cb-53e28cc38092-operator-scripts\") pod \"c1731159-7f86-4cbe-b9cb-53e28cc38092\" (UID: \"c1731159-7f86-4cbe-b9cb-53e28cc38092\") " Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.453871 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1731159-7f86-4cbe-b9cb-53e28cc38092-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c1731159-7f86-4cbe-b9cb-53e28cc38092" (UID: "c1731159-7f86-4cbe-b9cb-53e28cc38092"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.460461 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc952175-d111-4e8c-972d-e26cca11ba49-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "cc952175-d111-4e8c-972d-e26cca11ba49" (UID: "cc952175-d111-4e8c-972d-e26cca11ba49"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.465817 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1731159-7f86-4cbe-b9cb-53e28cc38092-kube-api-access-786xh" (OuterVolumeSpecName: "kube-api-access-786xh") pod "c1731159-7f86-4cbe-b9cb-53e28cc38092" (UID: "c1731159-7f86-4cbe-b9cb-53e28cc38092"). InnerVolumeSpecName "kube-api-access-786xh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.468533 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc952175-d111-4e8c-972d-e26cca11ba49-kube-api-access-2m29t" (OuterVolumeSpecName: "kube-api-access-2m29t") pod "cc952175-d111-4e8c-972d-e26cca11ba49" (UID: "cc952175-d111-4e8c-972d-e26cca11ba49"). InnerVolumeSpecName "kube-api-access-2m29t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.469643 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5f25-account-create-update-t9v5n" event={"ID":"c1731159-7f86-4cbe-b9cb-53e28cc38092","Type":"ContainerDied","Data":"5faf5915d32cee9be13c21b40165a86a817e59614d60cde4c00194198925cc1f"} Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.469677 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5faf5915d32cee9be13c21b40165a86a817e59614d60cde4c00194198925cc1f" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.469724 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5f25-account-create-update-t9v5n" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.474907 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-hvlz5" event={"ID":"ed73f0c1-2715-409f-9a4a-9da4fa8a3f48","Type":"ContainerDied","Data":"bc1fcdcf7b5d483a3c83ea270f2062ea95f1735fff6cfe07fc663a1dd1a95bb9"} Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.475130 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc1fcdcf7b5d483a3c83ea270f2062ea95f1735fff6cfe07fc663a1dd1a95bb9" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.475105 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-hvlz5" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.476559 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-k6ktf" event={"ID":"cc952175-d111-4e8c-972d-e26cca11ba49","Type":"ContainerDied","Data":"a4c94c4c83a134d1b09c5c1b41a4c1192348380f631f567df3d8c7274c9e957a"} Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.476658 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4c94c4c83a134d1b09c5c1b41a4c1192348380f631f567df3d8c7274c9e957a" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.476769 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-k6ktf" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.485526 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-0a6a-account-create-update-kvkqr" event={"ID":"eebc86bf-e7ee-4465-bd15-bef833ac0843","Type":"ContainerDied","Data":"a7b1a5751d7790154c64c4cba56830a10c7daa694f70e28037f4bed7bdabfc93"} Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.485564 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7b1a5751d7790154c64c4cba56830a10c7daa694f70e28037f4bed7bdabfc93" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.485698 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-0a6a-account-create-update-kvkqr" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.556577 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1731159-7f86-4cbe-b9cb-53e28cc38092-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.556606 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-786xh\" (UniqueName: \"kubernetes.io/projected/c1731159-7f86-4cbe-b9cb-53e28cc38092-kube-api-access-786xh\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.556616 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/cc952175-d111-4e8c-972d-e26cca11ba49-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.556629 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2m29t\" (UniqueName: \"kubernetes.io/projected/cc952175-d111-4e8c-972d-e26cca11ba49-kube-api-access-2m29t\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.891578 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-5fae-account-create-update-tdfsn"] Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.906158 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.965233 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-qnkx9"] Mar 13 09:33:42 crc kubenswrapper[4930]: I0313 09:33:42.986483 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-qnkx9"] Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.058462 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-ptkns"] Mar 13 09:33:43 crc kubenswrapper[4930]: E0313 09:33:43.059211 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eebc86bf-e7ee-4465-bd15-bef833ac0843" containerName="mariadb-account-create-update" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.059322 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="eebc86bf-e7ee-4465-bd15-bef833ac0843" containerName="mariadb-account-create-update" Mar 13 09:33:43 crc kubenswrapper[4930]: E0313 09:33:43.059424 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc952175-d111-4e8c-972d-e26cca11ba49" containerName="mariadb-database-create" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.059535 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc952175-d111-4e8c-972d-e26cca11ba49" containerName="mariadb-database-create" Mar 13 09:33:43 crc kubenswrapper[4930]: E0313 09:33:43.059634 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1731159-7f86-4cbe-b9cb-53e28cc38092" containerName="mariadb-account-create-update" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.059706 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1731159-7f86-4cbe-b9cb-53e28cc38092" containerName="mariadb-account-create-update" Mar 13 09:33:43 crc kubenswrapper[4930]: E0313 09:33:43.059786 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed73f0c1-2715-409f-9a4a-9da4fa8a3f48" containerName="mariadb-database-create" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.059857 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed73f0c1-2715-409f-9a4a-9da4fa8a3f48" containerName="mariadb-database-create" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.060351 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1731159-7f86-4cbe-b9cb-53e28cc38092" containerName="mariadb-account-create-update" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.060549 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc952175-d111-4e8c-972d-e26cca11ba49" containerName="mariadb-database-create" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.060813 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="eebc86bf-e7ee-4465-bd15-bef833ac0843" containerName="mariadb-account-create-update" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.060921 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed73f0c1-2715-409f-9a4a-9da4fa8a3f48" containerName="mariadb-database-create" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.062633 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-ptkns" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.075673 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-44qz9"] Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.079653 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.104833 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-8baa-account-create-update-b5w9f"] Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.128641 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-ptkns"] Mar 13 09:33:43 crc kubenswrapper[4930]: W0313 09:33:43.188889 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d7d0626_3583_4461_957d_d978d8f78ed4.slice/crio-230bda7378d84e35902c767b141162722bc66a2e4158d469b0b8d2c9171ee537 WatchSource:0}: Error finding container 230bda7378d84e35902c767b141162722bc66a2e4158d469b0b8d2c9171ee537: Status 404 returned error can't find the container with id 230bda7378d84e35902c767b141162722bc66a2e4158d469b0b8d2c9171ee537 Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.189138 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-bqwkt"] Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.191052 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-fghdj" podUID="2307735f-0b44-41ed-a622-74d62c05810c" containerName="ovn-controller" probeResult="failure" output=< Mar 13 09:33:43 crc kubenswrapper[4930]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 13 09:33:43 crc kubenswrapper[4930]: > Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.212727 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2cn7k\" (UniqueName: \"kubernetes.io/projected/902fb09a-cf93-4d3c-95d8-6098cb86b34b-kube-api-access-2cn7k\") pod \"root-account-create-update-ptkns\" (UID: \"902fb09a-cf93-4d3c-95d8-6098cb86b34b\") " pod="openstack/root-account-create-update-ptkns" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.213007 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/902fb09a-cf93-4d3c-95d8-6098cb86b34b-operator-scripts\") pod \"root-account-create-update-ptkns\" (UID: \"902fb09a-cf93-4d3c-95d8-6098cb86b34b\") " pod="openstack/root-account-create-update-ptkns" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.315892 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2cn7k\" (UniqueName: \"kubernetes.io/projected/902fb09a-cf93-4d3c-95d8-6098cb86b34b-kube-api-access-2cn7k\") pod \"root-account-create-update-ptkns\" (UID: \"902fb09a-cf93-4d3c-95d8-6098cb86b34b\") " pod="openstack/root-account-create-update-ptkns" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.315989 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/902fb09a-cf93-4d3c-95d8-6098cb86b34b-operator-scripts\") pod \"root-account-create-update-ptkns\" (UID: \"902fb09a-cf93-4d3c-95d8-6098cb86b34b\") " pod="openstack/root-account-create-update-ptkns" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.316895 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/902fb09a-cf93-4d3c-95d8-6098cb86b34b-operator-scripts\") pod \"root-account-create-update-ptkns\" (UID: \"902fb09a-cf93-4d3c-95d8-6098cb86b34b\") " pod="openstack/root-account-create-update-ptkns" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.350006 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2cn7k\" (UniqueName: \"kubernetes.io/projected/902fb09a-cf93-4d3c-95d8-6098cb86b34b-kube-api-access-2cn7k\") pod \"root-account-create-update-ptkns\" (UID: \"902fb09a-cf93-4d3c-95d8-6098cb86b34b\") " pod="openstack/root-account-create-update-ptkns" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.438460 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-ptkns" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.524179 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" event={"ID":"b0fe8888-5e4f-4bec-a5bf-554c209325ed","Type":"ContainerStarted","Data":"fa5304a13244c0dad9be995b80117f885c94a201f4733fc13ee62cc0a378d27b"} Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.524373 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" event={"ID":"b0fe8888-5e4f-4bec-a5bf-554c209325ed","Type":"ContainerStarted","Data":"c521a5e5dc12e9c4825284f9aa2d0d686a7cc637dcb37eaeb2ca44a7936a5a08"} Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.539213 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nqsgg" event={"ID":"2230cfae-6fbf-4df4-a249-622fe3c957e3","Type":"ContainerStarted","Data":"caef3f99b3bdbe53ee4e15b561a72966354fe0cc4a01acd6186956e320442c66"} Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.565793 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354","Type":"ContainerStarted","Data":"5d93f52f3a7b4038546d2005345570ab31cd4165cbad2ec0b14428854e754b01"} Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.566563 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.585227 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bqwkt" event={"ID":"0d7d0626-3583-4461-957d-d978d8f78ed4","Type":"ContainerStarted","Data":"01948896c7e83178c616925089deed8a34a3b95d918a9076a1e2364e6e254ddb"} Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.585279 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bqwkt" event={"ID":"0d7d0626-3583-4461-957d-d978d8f78ed4","Type":"ContainerStarted","Data":"230bda7378d84e35902c767b141162722bc66a2e4158d469b0b8d2c9171ee537"} Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.601256 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" podStartSLOduration=5.60122692 podStartE2EDuration="5.60122692s" podCreationTimestamp="2026-03-13 09:33:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:43.576991055 +0000 UTC m=+1264.326905732" watchObservedRunningTime="2026-03-13 09:33:43.60122692 +0000 UTC m=+1264.351141597" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.603502 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-44qz9" event={"ID":"451992e0-ca47-4af8-8b22-54e3c955b3aa","Type":"ContainerStarted","Data":"10b2cbb83b2d4a9450d5630b7c7bd2ee7ffb3c4f9cfb18964c42792c439ccf1a"} Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.603538 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-44qz9" event={"ID":"451992e0-ca47-4af8-8b22-54e3c955b3aa","Type":"ContainerStarted","Data":"4bd8c124d5ef75e426a706e5a8592ed06e8507da9a227f67f7cf8cc13cf316e8"} Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.647726 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-nqsgg" podStartSLOduration=8.262387946 podStartE2EDuration="13.647701726s" podCreationTimestamp="2026-03-13 09:33:30 +0000 UTC" firstStartedPulling="2026-03-13 09:33:36.938484332 +0000 UTC m=+1257.688399009" lastFinishedPulling="2026-03-13 09:33:42.323798112 +0000 UTC m=+1263.073712789" observedRunningTime="2026-03-13 09:33:43.6027715 +0000 UTC m=+1264.352686187" watchObservedRunningTime="2026-03-13 09:33:43.647701726 +0000 UTC m=+1264.397616393" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.671544 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.691266 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"897aa260-40c2-42d2-b59f-964a1a40116c","Type":"ContainerStarted","Data":"8fa8c46742e2c55c57c3f681b799c6123ebcb666f96bea9223dd69f96d660c17"} Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.692348 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.703823 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"2f69fa32-5b72-4f9b-9176-d43ee011605d","Type":"ContainerStarted","Data":"8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6"} Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.704284 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-1" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.708638 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=41.689714901 podStartE2EDuration="1m12.708618621s" podCreationTimestamp="2026-03-13 09:32:31 +0000 UTC" firstStartedPulling="2026-03-13 09:32:33.271120647 +0000 UTC m=+1194.021035324" lastFinishedPulling="2026-03-13 09:33:04.290024357 +0000 UTC m=+1225.039939044" observedRunningTime="2026-03-13 09:33:43.699517483 +0000 UTC m=+1264.449432160" watchObservedRunningTime="2026-03-13 09:33:43.708618621 +0000 UTC m=+1264.458533288" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.713999 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8baa-account-create-update-b5w9f" event={"ID":"f3efe40c-96ae-4b9e-a787-7781db5af486","Type":"ContainerStarted","Data":"50c93bba51bfa60c836f61ac84f93bcb53ca19ac740d2b512126ac639de9b8bf"} Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.714045 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8baa-account-create-update-b5w9f" event={"ID":"f3efe40c-96ae-4b9e-a787-7781db5af486","Type":"ContainerStarted","Data":"eb793f7e7fe831344a9053b9786b0356f243e9d8c117bcb8f8c6fc2f8330c5fd"} Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.725945 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=42.008718034 podStartE2EDuration="1m12.725929704s" podCreationTimestamp="2026-03-13 09:32:31 +0000 UTC" firstStartedPulling="2026-03-13 09:32:33.585204579 +0000 UTC m=+1194.335119256" lastFinishedPulling="2026-03-13 09:33:04.302416249 +0000 UTC m=+1225.052330926" observedRunningTime="2026-03-13 09:33:43.723007367 +0000 UTC m=+1264.472922044" watchObservedRunningTime="2026-03-13 09:33:43.725929704 +0000 UTC m=+1264.475844371" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.761337 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-bqwkt" podStartSLOduration=2.76131991 podStartE2EDuration="2.76131991s" podCreationTimestamp="2026-03-13 09:33:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:43.760324414 +0000 UTC m=+1264.510239111" watchObservedRunningTime="2026-03-13 09:33:43.76131991 +0000 UTC m=+1264.511234587" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.776606 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-openstack-db-create-44qz9" podStartSLOduration=5.77658868 podStartE2EDuration="5.77658868s" podCreationTimestamp="2026-03-13 09:33:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:43.774914506 +0000 UTC m=+1264.524829183" watchObservedRunningTime="2026-03-13 09:33:43.77658868 +0000 UTC m=+1264.526503347" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.878232 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-1" podStartSLOduration=44.607815626 podStartE2EDuration="1m12.87820672s" podCreationTimestamp="2026-03-13 09:32:31 +0000 UTC" firstStartedPulling="2026-03-13 09:32:33.864131977 +0000 UTC m=+1194.614046654" lastFinishedPulling="2026-03-13 09:33:02.134523061 +0000 UTC m=+1222.884437748" observedRunningTime="2026-03-13 09:33:43.856141192 +0000 UTC m=+1264.606055869" watchObservedRunningTime="2026-03-13 09:33:43.87820672 +0000 UTC m=+1264.628121387" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.879069 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-nkhpc"] Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.879360 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-nkhpc" podUID="af4031c0-0506-40a7-a5ae-2a3bb91a5521" containerName="dnsmasq-dns" containerID="cri-o://e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8" gracePeriod=10 Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.895168 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-8baa-account-create-update-b5w9f" podStartSLOduration=2.895151254 podStartE2EDuration="2.895151254s" podCreationTimestamp="2026-03-13 09:33:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:43.881340642 +0000 UTC m=+1264.631255329" watchObservedRunningTime="2026-03-13 09:33:43.895151254 +0000 UTC m=+1264.645065931" Mar 13 09:33:43 crc kubenswrapper[4930]: I0313 09:33:43.994577 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ff7901f-6245-437b-a7de-aa3e338fac88" path="/var/lib/kubelet/pods/5ff7901f-6245-437b-a7de-aa3e338fac88/volumes" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.077259 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-ptkns"] Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.426326 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.545477 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-ovsdbserver-nb\") pod \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.545719 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-klh98\" (UniqueName: \"kubernetes.io/projected/af4031c0-0506-40a7-a5ae-2a3bb91a5521-kube-api-access-klh98\") pod \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.545841 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-dns-svc\") pod \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.545891 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-config\") pod \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.545968 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-ovsdbserver-sb\") pod \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\" (UID: \"af4031c0-0506-40a7-a5ae-2a3bb91a5521\") " Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.575114 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af4031c0-0506-40a7-a5ae-2a3bb91a5521-kube-api-access-klh98" (OuterVolumeSpecName: "kube-api-access-klh98") pod "af4031c0-0506-40a7-a5ae-2a3bb91a5521" (UID: "af4031c0-0506-40a7-a5ae-2a3bb91a5521"). InnerVolumeSpecName "kube-api-access-klh98". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.603352 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-config" (OuterVolumeSpecName: "config") pod "af4031c0-0506-40a7-a5ae-2a3bb91a5521" (UID: "af4031c0-0506-40a7-a5ae-2a3bb91a5521"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.631023 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "af4031c0-0506-40a7-a5ae-2a3bb91a5521" (UID: "af4031c0-0506-40a7-a5ae-2a3bb91a5521"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.635819 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "af4031c0-0506-40a7-a5ae-2a3bb91a5521" (UID: "af4031c0-0506-40a7-a5ae-2a3bb91a5521"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.648436 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-klh98\" (UniqueName: \"kubernetes.io/projected/af4031c0-0506-40a7-a5ae-2a3bb91a5521-kube-api-access-klh98\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.648462 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.648484 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.648494 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.669088 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "af4031c0-0506-40a7-a5ae-2a3bb91a5521" (UID: "af4031c0-0506-40a7-a5ae-2a3bb91a5521"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.733409 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-ptkns" event={"ID":"902fb09a-cf93-4d3c-95d8-6098cb86b34b","Type":"ContainerStarted","Data":"0fd7a851ab276bfdb2cef98055c533b02ef0439cf4fc96152c66fafa107369e8"} Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.733486 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-ptkns" event={"ID":"902fb09a-cf93-4d3c-95d8-6098cb86b34b","Type":"ContainerStarted","Data":"458c04f7530c897534e9ad48411ea0bc3e9c1c72e9ba9403920284b3b405bc6e"} Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.750210 4930 generic.go:334] "Generic (PLEG): container finished" podID="af4031c0-0506-40a7-a5ae-2a3bb91a5521" containerID="e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8" exitCode=0 Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.750300 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-nkhpc" event={"ID":"af4031c0-0506-40a7-a5ae-2a3bb91a5521","Type":"ContainerDied","Data":"e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8"} Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.750370 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-nkhpc" event={"ID":"af4031c0-0506-40a7-a5ae-2a3bb91a5521","Type":"ContainerDied","Data":"9aa9e95d5a87075e73986c4fd0d098911715e9c0b467f2e42cbc7427d7b0bc53"} Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.750464 4930 scope.go:117] "RemoveContainer" containerID="e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.750853 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-nkhpc" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.759671 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af4031c0-0506-40a7-a5ae-2a3bb91a5521-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.760680 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-ptkns" podStartSLOduration=2.76066979 podStartE2EDuration="2.76066979s" podCreationTimestamp="2026-03-13 09:33:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:44.759834778 +0000 UTC m=+1265.509749465" watchObservedRunningTime="2026-03-13 09:33:44.76066979 +0000 UTC m=+1265.510584467" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.769217 4930 generic.go:334] "Generic (PLEG): container finished" podID="b0fe8888-5e4f-4bec-a5bf-554c209325ed" containerID="fa5304a13244c0dad9be995b80117f885c94a201f4733fc13ee62cc0a378d27b" exitCode=0 Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.769302 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" event={"ID":"b0fe8888-5e4f-4bec-a5bf-554c209325ed","Type":"ContainerDied","Data":"fa5304a13244c0dad9be995b80117f885c94a201f4733fc13ee62cc0a378d27b"} Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.784097 4930 generic.go:334] "Generic (PLEG): container finished" podID="0d7d0626-3583-4461-957d-d978d8f78ed4" containerID="01948896c7e83178c616925089deed8a34a3b95d918a9076a1e2364e6e254ddb" exitCode=0 Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.784175 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bqwkt" event={"ID":"0d7d0626-3583-4461-957d-d978d8f78ed4","Type":"ContainerDied","Data":"01948896c7e83178c616925089deed8a34a3b95d918a9076a1e2364e6e254ddb"} Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.791066 4930 generic.go:334] "Generic (PLEG): container finished" podID="451992e0-ca47-4af8-8b22-54e3c955b3aa" containerID="10b2cbb83b2d4a9450d5630b7c7bd2ee7ffb3c4f9cfb18964c42792c439ccf1a" exitCode=0 Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.791209 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-44qz9" event={"ID":"451992e0-ca47-4af8-8b22-54e3c955b3aa","Type":"ContainerDied","Data":"10b2cbb83b2d4a9450d5630b7c7bd2ee7ffb3c4f9cfb18964c42792c439ccf1a"} Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.804751 4930 scope.go:117] "RemoveContainer" containerID="3275277c91bfd35d1dafe4301b570a319af9112e60a19cde36448f4e939fba29" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.826064 4930 scope.go:117] "RemoveContainer" containerID="e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8" Mar 13 09:33:44 crc kubenswrapper[4930]: E0313 09:33:44.829791 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8\": container with ID starting with e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8 not found: ID does not exist" containerID="e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.829825 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8"} err="failed to get container status \"e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8\": rpc error: code = NotFound desc = could not find container \"e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8\": container with ID starting with e02cbae15c23edaf02cb32b4e8260eec33d66b3de04fce041999a85381a907c8 not found: ID does not exist" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.829849 4930 scope.go:117] "RemoveContainer" containerID="3275277c91bfd35d1dafe4301b570a319af9112e60a19cde36448f4e939fba29" Mar 13 09:33:44 crc kubenswrapper[4930]: E0313 09:33:44.832158 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3275277c91bfd35d1dafe4301b570a319af9112e60a19cde36448f4e939fba29\": container with ID starting with 3275277c91bfd35d1dafe4301b570a319af9112e60a19cde36448f4e939fba29 not found: ID does not exist" containerID="3275277c91bfd35d1dafe4301b570a319af9112e60a19cde36448f4e939fba29" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.832198 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3275277c91bfd35d1dafe4301b570a319af9112e60a19cde36448f4e939fba29"} err="failed to get container status \"3275277c91bfd35d1dafe4301b570a319af9112e60a19cde36448f4e939fba29\": rpc error: code = NotFound desc = could not find container \"3275277c91bfd35d1dafe4301b570a319af9112e60a19cde36448f4e939fba29\": container with ID starting with 3275277c91bfd35d1dafe4301b570a319af9112e60a19cde36448f4e939fba29 not found: ID does not exist" Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.838972 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-nkhpc"] Mar 13 09:33:44 crc kubenswrapper[4930]: I0313 09:33:44.853263 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-nkhpc"] Mar 13 09:33:45 crc kubenswrapper[4930]: I0313 09:33:45.380699 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:33:45 crc kubenswrapper[4930]: E0313 09:33:45.380945 4930 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 13 09:33:45 crc kubenswrapper[4930]: E0313 09:33:45.380992 4930 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 13 09:33:45 crc kubenswrapper[4930]: E0313 09:33:45.381058 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift podName:7b288113-babd-4a27-b397-cbdcbfdc195c nodeName:}" failed. No retries permitted until 2026-03-13 09:34:01.38103875 +0000 UTC m=+1282.130953427 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift") pod "swift-storage-0" (UID: "7b288113-babd-4a27-b397-cbdcbfdc195c") : configmap "swift-ring-files" not found Mar 13 09:33:45 crc kubenswrapper[4930]: I0313 09:33:45.825366 4930 generic.go:334] "Generic (PLEG): container finished" podID="902fb09a-cf93-4d3c-95d8-6098cb86b34b" containerID="0fd7a851ab276bfdb2cef98055c533b02ef0439cf4fc96152c66fafa107369e8" exitCode=0 Mar 13 09:33:45 crc kubenswrapper[4930]: I0313 09:33:45.825440 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-ptkns" event={"ID":"902fb09a-cf93-4d3c-95d8-6098cb86b34b","Type":"ContainerDied","Data":"0fd7a851ab276bfdb2cef98055c533b02ef0439cf4fc96152c66fafa107369e8"} Mar 13 09:33:45 crc kubenswrapper[4930]: I0313 09:33:45.829772 4930 generic.go:334] "Generic (PLEG): container finished" podID="f3efe40c-96ae-4b9e-a787-7781db5af486" containerID="50c93bba51bfa60c836f61ac84f93bcb53ca19ac740d2b512126ac639de9b8bf" exitCode=0 Mar 13 09:33:45 crc kubenswrapper[4930]: I0313 09:33:45.830523 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8baa-account-create-update-b5w9f" event={"ID":"f3efe40c-96ae-4b9e-a787-7781db5af486","Type":"ContainerDied","Data":"50c93bba51bfa60c836f61ac84f93bcb53ca19ac740d2b512126ac639de9b8bf"} Mar 13 09:33:45 crc kubenswrapper[4930]: I0313 09:33:45.984923 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af4031c0-0506-40a7-a5ae-2a3bb91a5521" path="/var/lib/kubelet/pods/af4031c0-0506-40a7-a5ae-2a3bb91a5521/volumes" Mar 13 09:33:46 crc kubenswrapper[4930]: I0313 09:33:46.842344 4930 generic.go:334] "Generic (PLEG): container finished" podID="a6623f09-f961-411a-9323-8cbf953c7148" containerID="96c39c7b929d20787c563592a28f65306475af49da6b2538643645d233ad69d4" exitCode=0 Mar 13 09:33:46 crc kubenswrapper[4930]: I0313 09:33:46.842573 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a6623f09-f961-411a-9323-8cbf953c7148","Type":"ContainerDied","Data":"96c39c7b929d20787c563592a28f65306475af49da6b2538643645d233ad69d4"} Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.831184 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-ptkns" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.844000 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.846535 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8baa-account-create-update-b5w9f" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.862809 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-44qz9" event={"ID":"451992e0-ca47-4af8-8b22-54e3c955b3aa","Type":"ContainerDied","Data":"4bd8c124d5ef75e426a706e5a8592ed06e8507da9a227f67f7cf8cc13cf316e8"} Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.862860 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bd8c124d5ef75e426a706e5a8592ed06e8507da9a227f67f7cf8cc13cf316e8" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.865681 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bqwkt" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.867041 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-ptkns" event={"ID":"902fb09a-cf93-4d3c-95d8-6098cb86b34b","Type":"ContainerDied","Data":"458c04f7530c897534e9ad48411ea0bc3e9c1c72e9ba9403920284b3b405bc6e"} Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.867073 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="458c04f7530c897534e9ad48411ea0bc3e9c1c72e9ba9403920284b3b405bc6e" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.867113 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-ptkns" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.895786 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-8baa-account-create-update-b5w9f" event={"ID":"f3efe40c-96ae-4b9e-a787-7781db5af486","Type":"ContainerDied","Data":"eb793f7e7fe831344a9053b9786b0356f243e9d8c117bcb8f8c6fc2f8330c5fd"} Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.895825 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb793f7e7fe831344a9053b9786b0356f243e9d8c117bcb8f8c6fc2f8330c5fd" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.895920 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-44qz9" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.895987 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-8baa-account-create-update-b5w9f" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.900045 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" event={"ID":"b0fe8888-5e4f-4bec-a5bf-554c209325ed","Type":"ContainerDied","Data":"c521a5e5dc12e9c4825284f9aa2d0d686a7cc637dcb37eaeb2ca44a7936a5a08"} Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.900123 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c521a5e5dc12e9c4825284f9aa2d0d686a7cc637dcb37eaeb2ca44a7936a5a08" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.900318 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-5fae-account-create-update-tdfsn" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.931147 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-bqwkt" event={"ID":"0d7d0626-3583-4461-957d-d978d8f78ed4","Type":"ContainerDied","Data":"230bda7378d84e35902c767b141162722bc66a2e4158d469b0b8d2c9171ee537"} Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.931183 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="230bda7378d84e35902c767b141162722bc66a2e4158d469b0b8d2c9171ee537" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.931233 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-bqwkt" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.944221 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d7d0626-3583-4461-957d-d978d8f78ed4-operator-scripts\") pod \"0d7d0626-3583-4461-957d-d978d8f78ed4\" (UID: \"0d7d0626-3583-4461-957d-d978d8f78ed4\") " Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.944261 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5n4v\" (UniqueName: \"kubernetes.io/projected/b0fe8888-5e4f-4bec-a5bf-554c209325ed-kube-api-access-w5n4v\") pod \"b0fe8888-5e4f-4bec-a5bf-554c209325ed\" (UID: \"b0fe8888-5e4f-4bec-a5bf-554c209325ed\") " Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.944443 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3efe40c-96ae-4b9e-a787-7781db5af486-operator-scripts\") pod \"f3efe40c-96ae-4b9e-a787-7781db5af486\" (UID: \"f3efe40c-96ae-4b9e-a787-7781db5af486\") " Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.944513 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/451992e0-ca47-4af8-8b22-54e3c955b3aa-operator-scripts\") pod \"451992e0-ca47-4af8-8b22-54e3c955b3aa\" (UID: \"451992e0-ca47-4af8-8b22-54e3c955b3aa\") " Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.944550 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cn7k\" (UniqueName: \"kubernetes.io/projected/902fb09a-cf93-4d3c-95d8-6098cb86b34b-kube-api-access-2cn7k\") pod \"902fb09a-cf93-4d3c-95d8-6098cb86b34b\" (UID: \"902fb09a-cf93-4d3c-95d8-6098cb86b34b\") " Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.944567 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ssg79\" (UniqueName: \"kubernetes.io/projected/0d7d0626-3583-4461-957d-d978d8f78ed4-kube-api-access-ssg79\") pod \"0d7d0626-3583-4461-957d-d978d8f78ed4\" (UID: \"0d7d0626-3583-4461-957d-d978d8f78ed4\") " Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.944619 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0fe8888-5e4f-4bec-a5bf-554c209325ed-operator-scripts\") pod \"b0fe8888-5e4f-4bec-a5bf-554c209325ed\" (UID: \"b0fe8888-5e4f-4bec-a5bf-554c209325ed\") " Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.944676 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r6fdz\" (UniqueName: \"kubernetes.io/projected/451992e0-ca47-4af8-8b22-54e3c955b3aa-kube-api-access-r6fdz\") pod \"451992e0-ca47-4af8-8b22-54e3c955b3aa\" (UID: \"451992e0-ca47-4af8-8b22-54e3c955b3aa\") " Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.944729 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/902fb09a-cf93-4d3c-95d8-6098cb86b34b-operator-scripts\") pod \"902fb09a-cf93-4d3c-95d8-6098cb86b34b\" (UID: \"902fb09a-cf93-4d3c-95d8-6098cb86b34b\") " Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.944749 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2cxr5\" (UniqueName: \"kubernetes.io/projected/f3efe40c-96ae-4b9e-a787-7781db5af486-kube-api-access-2cxr5\") pod \"f3efe40c-96ae-4b9e-a787-7781db5af486\" (UID: \"f3efe40c-96ae-4b9e-a787-7781db5af486\") " Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.946105 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b0fe8888-5e4f-4bec-a5bf-554c209325ed-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b0fe8888-5e4f-4bec-a5bf-554c209325ed" (UID: "b0fe8888-5e4f-4bec-a5bf-554c209325ed"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.946392 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3efe40c-96ae-4b9e-a787-7781db5af486-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f3efe40c-96ae-4b9e-a787-7781db5af486" (UID: "f3efe40c-96ae-4b9e-a787-7781db5af486"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.946652 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/451992e0-ca47-4af8-8b22-54e3c955b3aa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "451992e0-ca47-4af8-8b22-54e3c955b3aa" (UID: "451992e0-ca47-4af8-8b22-54e3c955b3aa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.947359 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0d7d0626-3583-4461-957d-d978d8f78ed4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0d7d0626-3583-4461-957d-d978d8f78ed4" (UID: "0d7d0626-3583-4461-957d-d978d8f78ed4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.947626 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/902fb09a-cf93-4d3c-95d8-6098cb86b34b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "902fb09a-cf93-4d3c-95d8-6098cb86b34b" (UID: "902fb09a-cf93-4d3c-95d8-6098cb86b34b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.952855 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b0fe8888-5e4f-4bec-a5bf-554c209325ed-kube-api-access-w5n4v" (OuterVolumeSpecName: "kube-api-access-w5n4v") pod "b0fe8888-5e4f-4bec-a5bf-554c209325ed" (UID: "b0fe8888-5e4f-4bec-a5bf-554c209325ed"). InnerVolumeSpecName "kube-api-access-w5n4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.953662 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/451992e0-ca47-4af8-8b22-54e3c955b3aa-kube-api-access-r6fdz" (OuterVolumeSpecName: "kube-api-access-r6fdz") pod "451992e0-ca47-4af8-8b22-54e3c955b3aa" (UID: "451992e0-ca47-4af8-8b22-54e3c955b3aa"). InnerVolumeSpecName "kube-api-access-r6fdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.954808 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d7d0626-3583-4461-957d-d978d8f78ed4-kube-api-access-ssg79" (OuterVolumeSpecName: "kube-api-access-ssg79") pod "0d7d0626-3583-4461-957d-d978d8f78ed4" (UID: "0d7d0626-3583-4461-957d-d978d8f78ed4"). InnerVolumeSpecName "kube-api-access-ssg79". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.955707 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3efe40c-96ae-4b9e-a787-7781db5af486-kube-api-access-2cxr5" (OuterVolumeSpecName: "kube-api-access-2cxr5") pod "f3efe40c-96ae-4b9e-a787-7781db5af486" (UID: "f3efe40c-96ae-4b9e-a787-7781db5af486"). InnerVolumeSpecName "kube-api-access-2cxr5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:47 crc kubenswrapper[4930]: I0313 09:33:47.966551 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/902fb09a-cf93-4d3c-95d8-6098cb86b34b-kube-api-access-2cn7k" (OuterVolumeSpecName: "kube-api-access-2cn7k") pod "902fb09a-cf93-4d3c-95d8-6098cb86b34b" (UID: "902fb09a-cf93-4d3c-95d8-6098cb86b34b"). InnerVolumeSpecName "kube-api-access-2cn7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.052113 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b0fe8888-5e4f-4bec-a5bf-554c209325ed-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.052566 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r6fdz\" (UniqueName: \"kubernetes.io/projected/451992e0-ca47-4af8-8b22-54e3c955b3aa-kube-api-access-r6fdz\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.052579 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/902fb09a-cf93-4d3c-95d8-6098cb86b34b-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.052588 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cxr5\" (UniqueName: \"kubernetes.io/projected/f3efe40c-96ae-4b9e-a787-7781db5af486-kube-api-access-2cxr5\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.052598 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5n4v\" (UniqueName: \"kubernetes.io/projected/b0fe8888-5e4f-4bec-a5bf-554c209325ed-kube-api-access-w5n4v\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.052607 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0d7d0626-3583-4461-957d-d978d8f78ed4-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.052616 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f3efe40c-96ae-4b9e-a787-7781db5af486-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.052625 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/451992e0-ca47-4af8-8b22-54e3c955b3aa-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.052634 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2cn7k\" (UniqueName: \"kubernetes.io/projected/902fb09a-cf93-4d3c-95d8-6098cb86b34b-kube-api-access-2cn7k\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.052642 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ssg79\" (UniqueName: \"kubernetes.io/projected/0d7d0626-3583-4461-957d-d978d8f78ed4-kube-api-access-ssg79\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.145761 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.146814 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.176772 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-fghdj" podUID="2307735f-0b44-41ed-a622-74d62c05810c" containerName="ovn-controller" probeResult="failure" output=< Mar 13 09:33:48 crc kubenswrapper[4930]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 13 09:33:48 crc kubenswrapper[4930]: > Mar 13 09:33:48 crc kubenswrapper[4930]: E0313 09:33:48.293193 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"prometheus\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openstack/prometheus-metric-storage-0" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.417733 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fghdj-config-46kdk"] Mar 13 09:33:48 crc kubenswrapper[4930]: E0313 09:33:48.418135 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="902fb09a-cf93-4d3c-95d8-6098cb86b34b" containerName="mariadb-account-create-update" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418150 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="902fb09a-cf93-4d3c-95d8-6098cb86b34b" containerName="mariadb-account-create-update" Mar 13 09:33:48 crc kubenswrapper[4930]: E0313 09:33:48.418166 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af4031c0-0506-40a7-a5ae-2a3bb91a5521" containerName="init" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418172 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="af4031c0-0506-40a7-a5ae-2a3bb91a5521" containerName="init" Mar 13 09:33:48 crc kubenswrapper[4930]: E0313 09:33:48.418182 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b0fe8888-5e4f-4bec-a5bf-554c209325ed" containerName="mariadb-account-create-update" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418189 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b0fe8888-5e4f-4bec-a5bf-554c209325ed" containerName="mariadb-account-create-update" Mar 13 09:33:48 crc kubenswrapper[4930]: E0313 09:33:48.418204 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="451992e0-ca47-4af8-8b22-54e3c955b3aa" containerName="mariadb-database-create" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418245 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="451992e0-ca47-4af8-8b22-54e3c955b3aa" containerName="mariadb-database-create" Mar 13 09:33:48 crc kubenswrapper[4930]: E0313 09:33:48.418257 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3efe40c-96ae-4b9e-a787-7781db5af486" containerName="mariadb-account-create-update" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418264 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3efe40c-96ae-4b9e-a787-7781db5af486" containerName="mariadb-account-create-update" Mar 13 09:33:48 crc kubenswrapper[4930]: E0313 09:33:48.418282 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af4031c0-0506-40a7-a5ae-2a3bb91a5521" containerName="dnsmasq-dns" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418289 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="af4031c0-0506-40a7-a5ae-2a3bb91a5521" containerName="dnsmasq-dns" Mar 13 09:33:48 crc kubenswrapper[4930]: E0313 09:33:48.418298 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d7d0626-3583-4461-957d-d978d8f78ed4" containerName="mariadb-database-create" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418305 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d7d0626-3583-4461-957d-d978d8f78ed4" containerName="mariadb-database-create" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418491 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="af4031c0-0506-40a7-a5ae-2a3bb91a5521" containerName="dnsmasq-dns" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418506 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="902fb09a-cf93-4d3c-95d8-6098cb86b34b" containerName="mariadb-account-create-update" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418514 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d7d0626-3583-4461-957d-d978d8f78ed4" containerName="mariadb-database-create" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418531 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3efe40c-96ae-4b9e-a787-7781db5af486" containerName="mariadb-account-create-update" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418540 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="451992e0-ca47-4af8-8b22-54e3c955b3aa" containerName="mariadb-database-create" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.418553 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b0fe8888-5e4f-4bec-a5bf-554c209325ed" containerName="mariadb-account-create-update" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.419177 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.421367 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.434773 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fghdj-config-46kdk"] Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.460196 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-scripts\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.460254 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-log-ovn\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.460290 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-run\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.460362 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-additional-scripts\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.460388 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wxvh\" (UniqueName: \"kubernetes.io/projected/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-kube-api-access-7wxvh\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.460410 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-run-ovn\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.562733 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-scripts\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.562817 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-log-ovn\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.562866 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-run\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.562953 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-additional-scripts\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.562990 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wxvh\" (UniqueName: \"kubernetes.io/projected/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-kube-api-access-7wxvh\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.563020 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-run-ovn\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.563563 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-run-ovn\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.563642 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-log-ovn\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.563693 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-run\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.564372 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-additional-scripts\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.564861 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-scripts\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.586667 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wxvh\" (UniqueName: \"kubernetes.io/projected/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-kube-api-access-7wxvh\") pod \"ovn-controller-fghdj-config-46kdk\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.750228 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.974358 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"267a25c3-20d9-4b6d-b097-d14b9534faa1","Type":"ContainerStarted","Data":"d523e85de62adeb9f18cd8d5c405896649ac61c40d0f00713e6a33213b02def4"} Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.985059 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a6623f09-f961-411a-9323-8cbf953c7148","Type":"ContainerStarted","Data":"b379c2c9b1a4f5934e943fc74b3b6eee20c34346290ab3b05975896af875e676"} Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.986245 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-44qz9" Mar 13 09:33:48 crc kubenswrapper[4930]: I0313 09:33:48.986277 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-2" Mar 13 09:33:49 crc kubenswrapper[4930]: I0313 09:33:49.115183 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-2" podStartSLOduration=-9223371958.739618 podStartE2EDuration="1m18.115156726s" podCreationTimestamp="2026-03-13 09:32:31 +0000 UTC" firstStartedPulling="2026-03-13 09:32:33.936659951 +0000 UTC m=+1194.686574628" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:49.076992147 +0000 UTC m=+1269.826906824" watchObservedRunningTime="2026-03-13 09:33:49.115156726 +0000 UTC m=+1269.865071403" Mar 13 09:33:49 crc kubenswrapper[4930]: I0313 09:33:49.339837 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fghdj-config-46kdk"] Mar 13 09:33:50 crc kubenswrapper[4930]: I0313 09:33:50.013172 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj-config-46kdk" event={"ID":"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3","Type":"ContainerStarted","Data":"544f3748e5dc2a3c11f5bae45b8a9e44d286062863d4c6013cca7699288a5348"} Mar 13 09:33:50 crc kubenswrapper[4930]: I0313 09:33:50.013520 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj-config-46kdk" event={"ID":"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3","Type":"ContainerStarted","Data":"88b8e9430ef6ed57cc6ebbdf8bf1179229f9778f14d9700d024b982484994d1b"} Mar 13 09:33:50 crc kubenswrapper[4930]: I0313 09:33:50.030866 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-fghdj-config-46kdk" podStartSLOduration=2.030848745 podStartE2EDuration="2.030848745s" podCreationTimestamp="2026-03-13 09:33:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:50.03067352 +0000 UTC m=+1270.780588197" watchObservedRunningTime="2026-03-13 09:33:50.030848745 +0000 UTC m=+1270.780763422" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.025263 4930 generic.go:334] "Generic (PLEG): container finished" podID="6bb8d7be-8ee2-4417-aedd-01e6cf0924e3" containerID="544f3748e5dc2a3c11f5bae45b8a9e44d286062863d4c6013cca7699288a5348" exitCode=0 Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.025315 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj-config-46kdk" event={"ID":"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3","Type":"ContainerDied","Data":"544f3748e5dc2a3c11f5bae45b8a9e44d286062863d4c6013cca7699288a5348"} Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.029514 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"267a25c3-20d9-4b6d-b097-d14b9534faa1","Type":"ContainerStarted","Data":"f51adb56f7ae4d763421fb6b107e55503b3d99f2cd6a3fecf2ad91a38012ff2b"} Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.080448 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=28.023376576 podStartE2EDuration="1m13.080417379s" podCreationTimestamp="2026-03-13 09:32:38 +0000 UTC" firstStartedPulling="2026-03-13 09:33:05.168109278 +0000 UTC m=+1225.918023955" lastFinishedPulling="2026-03-13 09:33:50.225150081 +0000 UTC m=+1270.975064758" observedRunningTime="2026-03-13 09:33:51.071978528 +0000 UTC m=+1271.821893205" watchObservedRunningTime="2026-03-13 09:33:51.080417379 +0000 UTC m=+1271.830332056" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.336363 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-g82c9"] Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.337925 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.348352 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gpm2r" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.348569 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.353334 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-g82c9"] Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.434909 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-db-sync-config-data\") pod \"glance-db-sync-g82c9\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.434979 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-combined-ca-bundle\") pod \"glance-db-sync-g82c9\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.435150 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-config-data\") pod \"glance-db-sync-g82c9\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.435206 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95f64\" (UniqueName: \"kubernetes.io/projected/bf10d12f-578b-400e-b480-60a7343bc344-kube-api-access-95f64\") pod \"glance-db-sync-g82c9\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.537208 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-db-sync-config-data\") pod \"glance-db-sync-g82c9\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.537280 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-combined-ca-bundle\") pod \"glance-db-sync-g82c9\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.537337 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-config-data\") pod \"glance-db-sync-g82c9\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.537385 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95f64\" (UniqueName: \"kubernetes.io/projected/bf10d12f-578b-400e-b480-60a7343bc344-kube-api-access-95f64\") pod \"glance-db-sync-g82c9\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.543904 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-db-sync-config-data\") pod \"glance-db-sync-g82c9\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.549104 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-combined-ca-bundle\") pod \"glance-db-sync-g82c9\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.555598 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-config-data\") pod \"glance-db-sync-g82c9\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.556071 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95f64\" (UniqueName: \"kubernetes.io/projected/bf10d12f-578b-400e-b480-60a7343bc344-kube-api-access-95f64\") pod \"glance-db-sync-g82c9\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:51 crc kubenswrapper[4930]: I0313 09:33:51.660759 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g82c9" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.379315 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-g82c9"] Mar 13 09:33:52 crc kubenswrapper[4930]: W0313 09:33:52.400288 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf10d12f_578b_400e_b480_60a7343bc344.slice/crio-d2a628d60e282a82ec857647464ea0b39ca58c7e8dfc03ce8c3c8ba0f8212941 WatchSource:0}: Error finding container d2a628d60e282a82ec857647464ea0b39ca58c7e8dfc03ce8c3c8ba0f8212941: Status 404 returned error can't find the container with id d2a628d60e282a82ec857647464ea0b39ca58c7e8dfc03ce8c3c8ba0f8212941 Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.456140 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.556428 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-scripts\") pod \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.557038 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-log-ovn\") pod \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.557126 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-additional-scripts\") pod \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.557159 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7wxvh\" (UniqueName: \"kubernetes.io/projected/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-kube-api-access-7wxvh\") pod \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.557166 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "6bb8d7be-8ee2-4417-aedd-01e6cf0924e3" (UID: "6bb8d7be-8ee2-4417-aedd-01e6cf0924e3"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.557235 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-run-ovn\") pod \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.557290 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-run\") pod \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\" (UID: \"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3\") " Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.557309 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "6bb8d7be-8ee2-4417-aedd-01e6cf0924e3" (UID: "6bb8d7be-8ee2-4417-aedd-01e6cf0924e3"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.557406 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-run" (OuterVolumeSpecName: "var-run") pod "6bb8d7be-8ee2-4417-aedd-01e6cf0924e3" (UID: "6bb8d7be-8ee2-4417-aedd-01e6cf0924e3"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.557810 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "6bb8d7be-8ee2-4417-aedd-01e6cf0924e3" (UID: "6bb8d7be-8ee2-4417-aedd-01e6cf0924e3"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.557991 4930 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.558014 4930 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.558028 4930 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.558037 4930 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.558664 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-scripts" (OuterVolumeSpecName: "scripts") pod "6bb8d7be-8ee2-4417-aedd-01e6cf0924e3" (UID: "6bb8d7be-8ee2-4417-aedd-01e6cf0924e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.566649 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-kube-api-access-7wxvh" (OuterVolumeSpecName: "kube-api-access-7wxvh") pod "6bb8d7be-8ee2-4417-aedd-01e6cf0924e3" (UID: "6bb8d7be-8ee2-4417-aedd-01e6cf0924e3"). InnerVolumeSpecName "kube-api-access-7wxvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.593643 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.134:5671: connect: connection refused" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.660574 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7wxvh\" (UniqueName: \"kubernetes.io/projected/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-kube-api-access-7wxvh\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:52 crc kubenswrapper[4930]: I0313 09:33:52.660650 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.022398 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="897aa260-40c2-42d2-b59f-964a1a40116c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.135:5671: connect: connection refused" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.051572 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g82c9" event={"ID":"bf10d12f-578b-400e-b480-60a7343bc344","Type":"ContainerStarted","Data":"d2a628d60e282a82ec857647464ea0b39ca58c7e8dfc03ce8c3c8ba0f8212941"} Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.054166 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj-config-46kdk" event={"ID":"6bb8d7be-8ee2-4417-aedd-01e6cf0924e3","Type":"ContainerDied","Data":"88b8e9430ef6ed57cc6ebbdf8bf1179229f9778f14d9700d024b982484994d1b"} Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.054211 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88b8e9430ef6ed57cc6ebbdf8bf1179229f9778f14d9700d024b982484994d1b" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.054236 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj-config-46kdk" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.085226 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-fghdj" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.141763 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-fghdj-config-46kdk"] Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.160537 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-fghdj-config-46kdk"] Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.161775 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="2f69fa32-5b72-4f9b-9176-d43ee011605d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.137:5671: connect: connection refused" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.218192 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fghdj-config-vrv7z"] Mar 13 09:33:53 crc kubenswrapper[4930]: E0313 09:33:53.218714 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6bb8d7be-8ee2-4417-aedd-01e6cf0924e3" containerName="ovn-config" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.218733 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6bb8d7be-8ee2-4417-aedd-01e6cf0924e3" containerName="ovn-config" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.218971 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6bb8d7be-8ee2-4417-aedd-01e6cf0924e3" containerName="ovn-config" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.219691 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.224904 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.233247 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fghdj-config-vrv7z"] Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.275348 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-run\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.275801 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd600dfb-9194-46a8-b272-0d248c6baf24-additional-scripts\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.276003 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bb29c\" (UniqueName: \"kubernetes.io/projected/cd600dfb-9194-46a8-b272-0d248c6baf24-kube-api-access-bb29c\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.276314 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd600dfb-9194-46a8-b272-0d248c6baf24-scripts\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.276390 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-run-ovn\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.276653 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-log-ovn\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.378146 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd600dfb-9194-46a8-b272-0d248c6baf24-additional-scripts\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.378406 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bb29c\" (UniqueName: \"kubernetes.io/projected/cd600dfb-9194-46a8-b272-0d248c6baf24-kube-api-access-bb29c\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.378541 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd600dfb-9194-46a8-b272-0d248c6baf24-scripts\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.378614 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-run-ovn\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.378730 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-log-ovn\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.378811 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-run\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.379190 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-run\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.379230 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-log-ovn\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.379229 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-run-ovn\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.379663 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd600dfb-9194-46a8-b272-0d248c6baf24-additional-scripts\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.381220 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd600dfb-9194-46a8-b272-0d248c6baf24-scripts\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.400478 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bb29c\" (UniqueName: \"kubernetes.io/projected/cd600dfb-9194-46a8-b272-0d248c6baf24-kube-api-access-bb29c\") pod \"ovn-controller-fghdj-config-vrv7z\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.544706 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.846897 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz"] Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.848741 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.874613 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz"] Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.891131 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7kb4\" (UniqueName: \"kubernetes.io/projected/4c20e5a2-621a-413c-8109-121f96f217b8-kube-api-access-f7kb4\") pod \"mysqld-exporter-openstack-cell1-db-create-b8gsz\" (UID: \"4c20e5a2-621a-413c-8109-121f96f217b8\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.891565 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c20e5a2-621a-413c-8109-121f96f217b8-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-b8gsz\" (UID: \"4c20e5a2-621a-413c-8109-121f96f217b8\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.987770 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6bb8d7be-8ee2-4417-aedd-01e6cf0924e3" path="/var/lib/kubelet/pods/6bb8d7be-8ee2-4417-aedd-01e6cf0924e3/volumes" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.993815 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7kb4\" (UniqueName: \"kubernetes.io/projected/4c20e5a2-621a-413c-8109-121f96f217b8-kube-api-access-f7kb4\") pod \"mysqld-exporter-openstack-cell1-db-create-b8gsz\" (UID: \"4c20e5a2-621a-413c-8109-121f96f217b8\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.993859 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c20e5a2-621a-413c-8109-121f96f217b8-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-b8gsz\" (UID: \"4c20e5a2-621a-413c-8109-121f96f217b8\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" Mar 13 09:33:53 crc kubenswrapper[4930]: I0313 09:33:53.994661 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c20e5a2-621a-413c-8109-121f96f217b8-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-b8gsz\" (UID: \"4c20e5a2-621a-413c-8109-121f96f217b8\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.029686 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7kb4\" (UniqueName: \"kubernetes.io/projected/4c20e5a2-621a-413c-8109-121f96f217b8-kube-api-access-f7kb4\") pod \"mysqld-exporter-openstack-cell1-db-create-b8gsz\" (UID: \"4c20e5a2-621a-413c-8109-121f96f217b8\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.058638 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0d69-account-create-update-rg6d8"] Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.059765 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" Mar 13 09:33:54 crc kubenswrapper[4930]: W0313 09:33:54.064669 4930 reflector.go:561] object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret": failed to list *v1.Secret: secrets "mysqld-exporter-openstack-cell1-db-secret" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Mar 13 09:33:54 crc kubenswrapper[4930]: E0313 09:33:54.064708 4930 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"mysqld-exporter-openstack-cell1-db-secret\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"mysqld-exporter-openstack-cell1-db-secret\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.086784 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c09afa02-80df-434d-999c-58e6787d0d0d","Type":"ContainerStarted","Data":"b907ed951ab6bda868766352a07cf4e153262e9e7f2449b4aeb4716c90b4d120"} Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.087030 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.090199 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0d69-account-create-update-rg6d8"] Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.160813 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.850655798 podStartE2EDuration="32.160787033s" podCreationTimestamp="2026-03-13 09:33:22 +0000 UTC" firstStartedPulling="2026-03-13 09:33:23.266388524 +0000 UTC m=+1244.016303201" lastFinishedPulling="2026-03-13 09:33:53.576519759 +0000 UTC m=+1274.326434436" observedRunningTime="2026-03-13 09:33:54.150483103 +0000 UTC m=+1274.900397800" watchObservedRunningTime="2026-03-13 09:33:54.160787033 +0000 UTC m=+1274.910701720" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.181291 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.201379 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/657e3e90-4b08-4df2-b773-1bbc1add41a7-operator-scripts\") pod \"mysqld-exporter-0d69-account-create-update-rg6d8\" (UID: \"657e3e90-4b08-4df2-b773-1bbc1add41a7\") " pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.201693 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htpfc\" (UniqueName: \"kubernetes.io/projected/657e3e90-4b08-4df2-b773-1bbc1add41a7-kube-api-access-htpfc\") pod \"mysqld-exporter-0d69-account-create-update-rg6d8\" (UID: \"657e3e90-4b08-4df2-b773-1bbc1add41a7\") " pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.244956 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fghdj-config-vrv7z"] Mar 13 09:33:54 crc kubenswrapper[4930]: W0313 09:33:54.289021 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd600dfb_9194_46a8_b272_0d248c6baf24.slice/crio-a89f21cec2c31be58d8eae454e33be892d6f5158b1c761bc62ca6b023e3281e8 WatchSource:0}: Error finding container a89f21cec2c31be58d8eae454e33be892d6f5158b1c761bc62ca6b023e3281e8: Status 404 returned error can't find the container with id a89f21cec2c31be58d8eae454e33be892d6f5158b1c761bc62ca6b023e3281e8 Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.304479 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/657e3e90-4b08-4df2-b773-1bbc1add41a7-operator-scripts\") pod \"mysqld-exporter-0d69-account-create-update-rg6d8\" (UID: \"657e3e90-4b08-4df2-b773-1bbc1add41a7\") " pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.304619 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htpfc\" (UniqueName: \"kubernetes.io/projected/657e3e90-4b08-4df2-b773-1bbc1add41a7-kube-api-access-htpfc\") pod \"mysqld-exporter-0d69-account-create-update-rg6d8\" (UID: \"657e3e90-4b08-4df2-b773-1bbc1add41a7\") " pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.305295 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/657e3e90-4b08-4df2-b773-1bbc1add41a7-operator-scripts\") pod \"mysqld-exporter-0d69-account-create-update-rg6d8\" (UID: \"657e3e90-4b08-4df2-b773-1bbc1add41a7\") " pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.340997 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htpfc\" (UniqueName: \"kubernetes.io/projected/657e3e90-4b08-4df2-b773-1bbc1add41a7-kube-api-access-htpfc\") pod \"mysqld-exporter-0d69-account-create-update-rg6d8\" (UID: \"657e3e90-4b08-4df2-b773-1bbc1add41a7\") " pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.405868 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.877729 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz"] Mar 13 09:33:54 crc kubenswrapper[4930]: W0313 09:33:54.905047 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c20e5a2_621a_413c_8109_121f96f217b8.slice/crio-369041c9b8c92e7b622e0d814ceb06a03b164fffef68aa11ed6f7fa00ecc11d4 WatchSource:0}: Error finding container 369041c9b8c92e7b622e0d814ceb06a03b164fffef68aa11ed6f7fa00ecc11d4: Status 404 returned error can't find the container with id 369041c9b8c92e7b622e0d814ceb06a03b164fffef68aa11ed6f7fa00ecc11d4 Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.924554 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Mar 13 09:33:54 crc kubenswrapper[4930]: I0313 09:33:54.996633 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0d69-account-create-update-rg6d8"] Mar 13 09:33:55 crc kubenswrapper[4930]: W0313 09:33:55.003290 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod657e3e90_4b08_4df2_b773_1bbc1add41a7.slice/crio-6e1ce02110004db1d48f73b1dedc83790a6ba99a1718ad0e46049f0129a48bb2 WatchSource:0}: Error finding container 6e1ce02110004db1d48f73b1dedc83790a6ba99a1718ad0e46049f0129a48bb2: Status 404 returned error can't find the container with id 6e1ce02110004db1d48f73b1dedc83790a6ba99a1718ad0e46049f0129a48bb2 Mar 13 09:33:55 crc kubenswrapper[4930]: I0313 09:33:55.075506 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Mar 13 09:33:55 crc kubenswrapper[4930]: I0313 09:33:55.075547 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 13 09:33:55 crc kubenswrapper[4930]: I0313 09:33:55.080572 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Mar 13 09:33:55 crc kubenswrapper[4930]: I0313 09:33:55.108732 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" event={"ID":"657e3e90-4b08-4df2-b773-1bbc1add41a7","Type":"ContainerStarted","Data":"6e1ce02110004db1d48f73b1dedc83790a6ba99a1718ad0e46049f0129a48bb2"} Mar 13 09:33:55 crc kubenswrapper[4930]: I0313 09:33:55.122249 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" event={"ID":"4c20e5a2-621a-413c-8109-121f96f217b8","Type":"ContainerStarted","Data":"369041c9b8c92e7b622e0d814ceb06a03b164fffef68aa11ed6f7fa00ecc11d4"} Mar 13 09:33:55 crc kubenswrapper[4930]: I0313 09:33:55.130365 4930 generic.go:334] "Generic (PLEG): container finished" podID="2230cfae-6fbf-4df4-a249-622fe3c957e3" containerID="caef3f99b3bdbe53ee4e15b561a72966354fe0cc4a01acd6186956e320442c66" exitCode=0 Mar 13 09:33:55 crc kubenswrapper[4930]: I0313 09:33:55.130469 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nqsgg" event={"ID":"2230cfae-6fbf-4df4-a249-622fe3c957e3","Type":"ContainerDied","Data":"caef3f99b3bdbe53ee4e15b561a72966354fe0cc4a01acd6186956e320442c66"} Mar 13 09:33:55 crc kubenswrapper[4930]: I0313 09:33:55.145635 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj-config-vrv7z" event={"ID":"cd600dfb-9194-46a8-b272-0d248c6baf24","Type":"ContainerStarted","Data":"0ff24caf2459e3a24991ce35ffc0618b23e74e2db943bd600c2e1db6808a42da"} Mar 13 09:33:55 crc kubenswrapper[4930]: I0313 09:33:55.145675 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj-config-vrv7z" event={"ID":"cd600dfb-9194-46a8-b272-0d248c6baf24","Type":"ContainerStarted","Data":"a89f21cec2c31be58d8eae454e33be892d6f5158b1c761bc62ca6b023e3281e8"} Mar 13 09:33:55 crc kubenswrapper[4930]: I0313 09:33:55.147368 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Mar 13 09:33:55 crc kubenswrapper[4930]: I0313 09:33:55.265405 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-fghdj-config-vrv7z" podStartSLOduration=2.265385268 podStartE2EDuration="2.265385268s" podCreationTimestamp="2026-03-13 09:33:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:33:55.252335846 +0000 UTC m=+1276.002250533" watchObservedRunningTime="2026-03-13 09:33:55.265385268 +0000 UTC m=+1276.015299945" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.156558 4930 generic.go:334] "Generic (PLEG): container finished" podID="657e3e90-4b08-4df2-b773-1bbc1add41a7" containerID="c45ac4a97f6a89041787ac9d397f7972f50c469a10f0e15b010101bf8dc2831b" exitCode=0 Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.156663 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" event={"ID":"657e3e90-4b08-4df2-b773-1bbc1add41a7","Type":"ContainerDied","Data":"c45ac4a97f6a89041787ac9d397f7972f50c469a10f0e15b010101bf8dc2831b"} Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.160058 4930 generic.go:334] "Generic (PLEG): container finished" podID="4c20e5a2-621a-413c-8109-121f96f217b8" containerID="2ee5212ad57c16330e57fe3aed5020f51bdd1d7af3441124b4fbc62930021d66" exitCode=0 Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.160144 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" event={"ID":"4c20e5a2-621a-413c-8109-121f96f217b8","Type":"ContainerDied","Data":"2ee5212ad57c16330e57fe3aed5020f51bdd1d7af3441124b4fbc62930021d66"} Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.163511 4930 generic.go:334] "Generic (PLEG): container finished" podID="cd600dfb-9194-46a8-b272-0d248c6baf24" containerID="0ff24caf2459e3a24991ce35ffc0618b23e74e2db943bd600c2e1db6808a42da" exitCode=0 Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.163625 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj-config-vrv7z" event={"ID":"cd600dfb-9194-46a8-b272-0d248c6baf24","Type":"ContainerDied","Data":"0ff24caf2459e3a24991ce35ffc0618b23e74e2db943bd600c2e1db6808a42da"} Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.621892 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.760841 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2230cfae-6fbf-4df4-a249-622fe3c957e3-etc-swift\") pod \"2230cfae-6fbf-4df4-a249-622fe3c957e3\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.760983 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2230cfae-6fbf-4df4-a249-622fe3c957e3-ring-data-devices\") pod \"2230cfae-6fbf-4df4-a249-622fe3c957e3\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.761011 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-dispersionconf\") pod \"2230cfae-6fbf-4df4-a249-622fe3c957e3\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.761033 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2230cfae-6fbf-4df4-a249-622fe3c957e3-scripts\") pod \"2230cfae-6fbf-4df4-a249-622fe3c957e3\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.761129 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-combined-ca-bundle\") pod \"2230cfae-6fbf-4df4-a249-622fe3c957e3\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.761205 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-swiftconf\") pod \"2230cfae-6fbf-4df4-a249-622fe3c957e3\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.761246 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xbqgl\" (UniqueName: \"kubernetes.io/projected/2230cfae-6fbf-4df4-a249-622fe3c957e3-kube-api-access-xbqgl\") pod \"2230cfae-6fbf-4df4-a249-622fe3c957e3\" (UID: \"2230cfae-6fbf-4df4-a249-622fe3c957e3\") " Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.764688 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2230cfae-6fbf-4df4-a249-622fe3c957e3-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "2230cfae-6fbf-4df4-a249-622fe3c957e3" (UID: "2230cfae-6fbf-4df4-a249-622fe3c957e3"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.765325 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2230cfae-6fbf-4df4-a249-622fe3c957e3-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "2230cfae-6fbf-4df4-a249-622fe3c957e3" (UID: "2230cfae-6fbf-4df4-a249-622fe3c957e3"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.777515 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2230cfae-6fbf-4df4-a249-622fe3c957e3-kube-api-access-xbqgl" (OuterVolumeSpecName: "kube-api-access-xbqgl") pod "2230cfae-6fbf-4df4-a249-622fe3c957e3" (UID: "2230cfae-6fbf-4df4-a249-622fe3c957e3"). InnerVolumeSpecName "kube-api-access-xbqgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.796331 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "2230cfae-6fbf-4df4-a249-622fe3c957e3" (UID: "2230cfae-6fbf-4df4-a249-622fe3c957e3"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.800061 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2230cfae-6fbf-4df4-a249-622fe3c957e3" (UID: "2230cfae-6fbf-4df4-a249-622fe3c957e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.800275 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "2230cfae-6fbf-4df4-a249-622fe3c957e3" (UID: "2230cfae-6fbf-4df4-a249-622fe3c957e3"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.806388 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2230cfae-6fbf-4df4-a249-622fe3c957e3-scripts" (OuterVolumeSpecName: "scripts") pod "2230cfae-6fbf-4df4-a249-622fe3c957e3" (UID: "2230cfae-6fbf-4df4-a249-622fe3c957e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.863620 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.863659 4930 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.863669 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xbqgl\" (UniqueName: \"kubernetes.io/projected/2230cfae-6fbf-4df4-a249-622fe3c957e3-kube-api-access-xbqgl\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.863680 4930 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/2230cfae-6fbf-4df4-a249-622fe3c957e3-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.863689 4930 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/2230cfae-6fbf-4df4-a249-622fe3c957e3-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.863701 4930 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/2230cfae-6fbf-4df4-a249-622fe3c957e3-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:56 crc kubenswrapper[4930]: I0313 09:33:56.863710 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2230cfae-6fbf-4df4-a249-622fe3c957e3-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:57 crc kubenswrapper[4930]: I0313 09:33:57.185668 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-nqsgg" Mar 13 09:33:57 crc kubenswrapper[4930]: I0313 09:33:57.185757 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-nqsgg" event={"ID":"2230cfae-6fbf-4df4-a249-622fe3c957e3","Type":"ContainerDied","Data":"416aef5b6b5eeda5dd38a6d87a26acba20e41cb21b52f85ecde82d776429f6ad"} Mar 13 09:33:57 crc kubenswrapper[4930]: I0313 09:33:57.185793 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="416aef5b6b5eeda5dd38a6d87a26acba20e41cb21b52f85ecde82d776429f6ad" Mar 13 09:33:57 crc kubenswrapper[4930]: I0313 09:33:57.672906 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" Mar 13 09:33:57 crc kubenswrapper[4930]: I0313 09:33:57.821734 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c20e5a2-621a-413c-8109-121f96f217b8-operator-scripts\") pod \"4c20e5a2-621a-413c-8109-121f96f217b8\" (UID: \"4c20e5a2-621a-413c-8109-121f96f217b8\") " Mar 13 09:33:57 crc kubenswrapper[4930]: I0313 09:33:57.821964 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f7kb4\" (UniqueName: \"kubernetes.io/projected/4c20e5a2-621a-413c-8109-121f96f217b8-kube-api-access-f7kb4\") pod \"4c20e5a2-621a-413c-8109-121f96f217b8\" (UID: \"4c20e5a2-621a-413c-8109-121f96f217b8\") " Mar 13 09:33:57 crc kubenswrapper[4930]: I0313 09:33:57.823198 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c20e5a2-621a-413c-8109-121f96f217b8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4c20e5a2-621a-413c-8109-121f96f217b8" (UID: "4c20e5a2-621a-413c-8109-121f96f217b8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:57 crc kubenswrapper[4930]: I0313 09:33:57.838530 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c20e5a2-621a-413c-8109-121f96f217b8-kube-api-access-f7kb4" (OuterVolumeSpecName: "kube-api-access-f7kb4") pod "4c20e5a2-621a-413c-8109-121f96f217b8" (UID: "4c20e5a2-621a-413c-8109-121f96f217b8"). InnerVolumeSpecName "kube-api-access-f7kb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:57 crc kubenswrapper[4930]: I0313 09:33:57.912223 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:57 crc kubenswrapper[4930]: I0313 09:33:57.919842 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" Mar 13 09:33:57 crc kubenswrapper[4930]: I0313 09:33:57.924897 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4c20e5a2-621a-413c-8109-121f96f217b8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:57 crc kubenswrapper[4930]: I0313 09:33:57.924946 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f7kb4\" (UniqueName: \"kubernetes.io/projected/4c20e5a2-621a-413c-8109-121f96f217b8-kube-api-access-f7kb4\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.026151 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-run\") pod \"cd600dfb-9194-46a8-b272-0d248c6baf24\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.026664 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-run" (OuterVolumeSpecName: "var-run") pod "cd600dfb-9194-46a8-b272-0d248c6baf24" (UID: "cd600dfb-9194-46a8-b272-0d248c6baf24"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.026724 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-run-ovn\") pod \"cd600dfb-9194-46a8-b272-0d248c6baf24\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.026757 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bb29c\" (UniqueName: \"kubernetes.io/projected/cd600dfb-9194-46a8-b272-0d248c6baf24-kube-api-access-bb29c\") pod \"cd600dfb-9194-46a8-b272-0d248c6baf24\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.026836 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/657e3e90-4b08-4df2-b773-1bbc1add41a7-operator-scripts\") pod \"657e3e90-4b08-4df2-b773-1bbc1add41a7\" (UID: \"657e3e90-4b08-4df2-b773-1bbc1add41a7\") " Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.026923 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-log-ovn\") pod \"cd600dfb-9194-46a8-b272-0d248c6baf24\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.026952 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd600dfb-9194-46a8-b272-0d248c6baf24-additional-scripts\") pod \"cd600dfb-9194-46a8-b272-0d248c6baf24\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.026982 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd600dfb-9194-46a8-b272-0d248c6baf24-scripts\") pod \"cd600dfb-9194-46a8-b272-0d248c6baf24\" (UID: \"cd600dfb-9194-46a8-b272-0d248c6baf24\") " Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.027094 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htpfc\" (UniqueName: \"kubernetes.io/projected/657e3e90-4b08-4df2-b773-1bbc1add41a7-kube-api-access-htpfc\") pod \"657e3e90-4b08-4df2-b773-1bbc1add41a7\" (UID: \"657e3e90-4b08-4df2-b773-1bbc1add41a7\") " Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.027867 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/657e3e90-4b08-4df2-b773-1bbc1add41a7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "657e3e90-4b08-4df2-b773-1bbc1add41a7" (UID: "657e3e90-4b08-4df2-b773-1bbc1add41a7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.027907 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "cd600dfb-9194-46a8-b272-0d248c6baf24" (UID: "cd600dfb-9194-46a8-b272-0d248c6baf24"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.028947 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd600dfb-9194-46a8-b272-0d248c6baf24-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "cd600dfb-9194-46a8-b272-0d248c6baf24" (UID: "cd600dfb-9194-46a8-b272-0d248c6baf24"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.028973 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "cd600dfb-9194-46a8-b272-0d248c6baf24" (UID: "cd600dfb-9194-46a8-b272-0d248c6baf24"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.029669 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd600dfb-9194-46a8-b272-0d248c6baf24-scripts" (OuterVolumeSpecName: "scripts") pod "cd600dfb-9194-46a8-b272-0d248c6baf24" (UID: "cd600dfb-9194-46a8-b272-0d248c6baf24"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.053688 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd600dfb-9194-46a8-b272-0d248c6baf24-kube-api-access-bb29c" (OuterVolumeSpecName: "kube-api-access-bb29c") pod "cd600dfb-9194-46a8-b272-0d248c6baf24" (UID: "cd600dfb-9194-46a8-b272-0d248c6baf24"). InnerVolumeSpecName "kube-api-access-bb29c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.054824 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/657e3e90-4b08-4df2-b773-1bbc1add41a7-kube-api-access-htpfc" (OuterVolumeSpecName: "kube-api-access-htpfc") pod "657e3e90-4b08-4df2-b773-1bbc1add41a7" (UID: "657e3e90-4b08-4df2-b773-1bbc1add41a7"). InnerVolumeSpecName "kube-api-access-htpfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.076171 4930 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.076212 4930 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/cd600dfb-9194-46a8-b272-0d248c6baf24-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.076228 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd600dfb-9194-46a8-b272-0d248c6baf24-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.076242 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htpfc\" (UniqueName: \"kubernetes.io/projected/657e3e90-4b08-4df2-b773-1bbc1add41a7-kube-api-access-htpfc\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.076251 4930 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.076258 4930 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cd600dfb-9194-46a8-b272-0d248c6baf24-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.076268 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bb29c\" (UniqueName: \"kubernetes.io/projected/cd600dfb-9194-46a8-b272-0d248c6baf24-kube-api-access-bb29c\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.076276 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/657e3e90-4b08-4df2-b773-1bbc1add41a7-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.088817 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.089077 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="config-reloader" containerID="cri-o://9aaa51fb4611379e8c17b3648f7eeeccd410a16f268c69209f0374f1300e0f58" gracePeriod=600 Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.090593 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="prometheus" containerID="cri-o://f51adb56f7ae4d763421fb6b107e55503b3d99f2cd6a3fecf2ad91a38012ff2b" gracePeriod=600 Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.090768 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="thanos-sidecar" containerID="cri-o://d523e85de62adeb9f18cd8d5c405896649ac61c40d0f00713e6a33213b02def4" gracePeriod=600 Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.226382 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" event={"ID":"4c20e5a2-621a-413c-8109-121f96f217b8","Type":"ContainerDied","Data":"369041c9b8c92e7b622e0d814ceb06a03b164fffef68aa11ed6f7fa00ecc11d4"} Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.226424 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="369041c9b8c92e7b622e0d814ceb06a03b164fffef68aa11ed6f7fa00ecc11d4" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.226487 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.253798 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj-config-vrv7z" event={"ID":"cd600dfb-9194-46a8-b272-0d248c6baf24","Type":"ContainerDied","Data":"a89f21cec2c31be58d8eae454e33be892d6f5158b1c761bc62ca6b023e3281e8"} Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.253848 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a89f21cec2c31be58d8eae454e33be892d6f5158b1c761bc62ca6b023e3281e8" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.253924 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj-config-vrv7z" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.276019 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" event={"ID":"657e3e90-4b08-4df2-b773-1bbc1add41a7","Type":"ContainerDied","Data":"6e1ce02110004db1d48f73b1dedc83790a6ba99a1718ad0e46049f0129a48bb2"} Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.276071 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e1ce02110004db1d48f73b1dedc83790a6ba99a1718ad0e46049f0129a48bb2" Mar 13 09:33:58 crc kubenswrapper[4930]: I0313 09:33:58.276147 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0d69-account-create-update-rg6d8" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.008529 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-fghdj-config-vrv7z"] Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.021028 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-fghdj-config-vrv7z"] Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.118092 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-fghdj-config-jppqn"] Mar 13 09:33:59 crc kubenswrapper[4930]: E0313 09:33:59.118932 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="657e3e90-4b08-4df2-b773-1bbc1add41a7" containerName="mariadb-account-create-update" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.118962 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="657e3e90-4b08-4df2-b773-1bbc1add41a7" containerName="mariadb-account-create-update" Mar 13 09:33:59 crc kubenswrapper[4930]: E0313 09:33:59.118983 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2230cfae-6fbf-4df4-a249-622fe3c957e3" containerName="swift-ring-rebalance" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.118994 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2230cfae-6fbf-4df4-a249-622fe3c957e3" containerName="swift-ring-rebalance" Mar 13 09:33:59 crc kubenswrapper[4930]: E0313 09:33:59.119006 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd600dfb-9194-46a8-b272-0d248c6baf24" containerName="ovn-config" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.119015 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd600dfb-9194-46a8-b272-0d248c6baf24" containerName="ovn-config" Mar 13 09:33:59 crc kubenswrapper[4930]: E0313 09:33:59.119035 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c20e5a2-621a-413c-8109-121f96f217b8" containerName="mariadb-database-create" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.119043 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c20e5a2-621a-413c-8109-121f96f217b8" containerName="mariadb-database-create" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.119301 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="2230cfae-6fbf-4df4-a249-622fe3c957e3" containerName="swift-ring-rebalance" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.119337 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="657e3e90-4b08-4df2-b773-1bbc1add41a7" containerName="mariadb-account-create-update" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.119358 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c20e5a2-621a-413c-8109-121f96f217b8" containerName="mariadb-database-create" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.119371 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd600dfb-9194-46a8-b272-0d248c6baf24" containerName="ovn-config" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.120202 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.123365 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.136330 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fghdj-config-jppqn"] Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.204481 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cj6fg\" (UniqueName: \"kubernetes.io/projected/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-kube-api-access-cj6fg\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.204575 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-run-ovn\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.204610 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-additional-scripts\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.204672 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-log-ovn\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.204708 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-run\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.204731 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-scripts\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.210175 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.211502 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.217084 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.221773 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.290225 4930 generic.go:334] "Generic (PLEG): container finished" podID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerID="f51adb56f7ae4d763421fb6b107e55503b3d99f2cd6a3fecf2ad91a38012ff2b" exitCode=0 Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.290264 4930 generic.go:334] "Generic (PLEG): container finished" podID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerID="d523e85de62adeb9f18cd8d5c405896649ac61c40d0f00713e6a33213b02def4" exitCode=0 Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.290276 4930 generic.go:334] "Generic (PLEG): container finished" podID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerID="9aaa51fb4611379e8c17b3648f7eeeccd410a16f268c69209f0374f1300e0f58" exitCode=0 Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.290298 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"267a25c3-20d9-4b6d-b097-d14b9534faa1","Type":"ContainerDied","Data":"f51adb56f7ae4d763421fb6b107e55503b3d99f2cd6a3fecf2ad91a38012ff2b"} Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.290329 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"267a25c3-20d9-4b6d-b097-d14b9534faa1","Type":"ContainerDied","Data":"d523e85de62adeb9f18cd8d5c405896649ac61c40d0f00713e6a33213b02def4"} Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.290345 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"267a25c3-20d9-4b6d-b097-d14b9534faa1","Type":"ContainerDied","Data":"9aaa51fb4611379e8c17b3648f7eeeccd410a16f268c69209f0374f1300e0f58"} Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.306938 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-log-ovn\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.307004 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-run\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.307042 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-scripts\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.307109 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cj6fg\" (UniqueName: \"kubernetes.io/projected/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-kube-api-access-cj6fg\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.307206 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97bec685-87be-4c44-8678-189f13ffce7c-config-data\") pod \"mysqld-exporter-0\" (UID: \"97bec685-87be-4c44-8678-189f13ffce7c\") " pod="openstack/mysqld-exporter-0" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.307231 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzdb6\" (UniqueName: \"kubernetes.io/projected/97bec685-87be-4c44-8678-189f13ffce7c-kube-api-access-mzdb6\") pod \"mysqld-exporter-0\" (UID: \"97bec685-87be-4c44-8678-189f13ffce7c\") " pod="openstack/mysqld-exporter-0" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.307257 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bec685-87be-4c44-8678-189f13ffce7c-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"97bec685-87be-4c44-8678-189f13ffce7c\") " pod="openstack/mysqld-exporter-0" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.307283 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-run\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.307243 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-log-ovn\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.307411 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-run-ovn\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.307515 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-run-ovn\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.307523 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-additional-scripts\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.308769 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-additional-scripts\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.315358 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-scripts\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.328154 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cj6fg\" (UniqueName: \"kubernetes.io/projected/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-kube-api-access-cj6fg\") pod \"ovn-controller-fghdj-config-jppqn\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.420698 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97bec685-87be-4c44-8678-189f13ffce7c-config-data\") pod \"mysqld-exporter-0\" (UID: \"97bec685-87be-4c44-8678-189f13ffce7c\") " pod="openstack/mysqld-exporter-0" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.420742 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzdb6\" (UniqueName: \"kubernetes.io/projected/97bec685-87be-4c44-8678-189f13ffce7c-kube-api-access-mzdb6\") pod \"mysqld-exporter-0\" (UID: \"97bec685-87be-4c44-8678-189f13ffce7c\") " pod="openstack/mysqld-exporter-0" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.420763 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bec685-87be-4c44-8678-189f13ffce7c-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"97bec685-87be-4c44-8678-189f13ffce7c\") " pod="openstack/mysqld-exporter-0" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.434095 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bec685-87be-4c44-8678-189f13ffce7c-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"97bec685-87be-4c44-8678-189f13ffce7c\") " pod="openstack/mysqld-exporter-0" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.438294 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97bec685-87be-4c44-8678-189f13ffce7c-config-data\") pod \"mysqld-exporter-0\" (UID: \"97bec685-87be-4c44-8678-189f13ffce7c\") " pod="openstack/mysqld-exporter-0" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.449049 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzdb6\" (UniqueName: \"kubernetes.io/projected/97bec685-87be-4c44-8678-189f13ffce7c-kube-api-access-mzdb6\") pod \"mysqld-exporter-0\" (UID: \"97bec685-87be-4c44-8678-189f13ffce7c\") " pod="openstack/mysqld-exporter-0" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.449514 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.561928 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.728191 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.832932 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-2\") pod \"267a25c3-20d9-4b6d-b097-d14b9534faa1\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.833131 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\") pod \"267a25c3-20d9-4b6d-b097-d14b9534faa1\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.833277 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/267a25c3-20d9-4b6d-b097-d14b9534faa1-tls-assets\") pod \"267a25c3-20d9-4b6d-b097-d14b9534faa1\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.833367 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-thanos-prometheus-http-client-file\") pod \"267a25c3-20d9-4b6d-b097-d14b9534faa1\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.833756 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-1\") pod \"267a25c3-20d9-4b6d-b097-d14b9534faa1\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.833913 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9mvlv\" (UniqueName: \"kubernetes.io/projected/267a25c3-20d9-4b6d-b097-d14b9534faa1-kube-api-access-9mvlv\") pod \"267a25c3-20d9-4b6d-b097-d14b9534faa1\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.833958 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-config\") pod \"267a25c3-20d9-4b6d-b097-d14b9534faa1\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.833988 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-web-config\") pod \"267a25c3-20d9-4b6d-b097-d14b9534faa1\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.833992 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-2" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-2") pod "267a25c3-20d9-4b6d-b097-d14b9534faa1" (UID: "267a25c3-20d9-4b6d-b097-d14b9534faa1"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-2". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.834020 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-0\") pod \"267a25c3-20d9-4b6d-b097-d14b9534faa1\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.834068 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/267a25c3-20d9-4b6d-b097-d14b9534faa1-config-out\") pod \"267a25c3-20d9-4b6d-b097-d14b9534faa1\" (UID: \"267a25c3-20d9-4b6d-b097-d14b9534faa1\") " Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.834375 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-1" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-1") pod "267a25c3-20d9-4b6d-b097-d14b9534faa1" (UID: "267a25c3-20d9-4b6d-b097-d14b9534faa1"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-1". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.835212 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "267a25c3-20d9-4b6d-b097-d14b9534faa1" (UID: "267a25c3-20d9-4b6d-b097-d14b9534faa1"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.835498 4930 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-2\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.835524 4930 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-1\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.835537 4930 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/267a25c3-20d9-4b6d-b097-d14b9534faa1-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.842149 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/267a25c3-20d9-4b6d-b097-d14b9534faa1-config-out" (OuterVolumeSpecName: "config-out") pod "267a25c3-20d9-4b6d-b097-d14b9534faa1" (UID: "267a25c3-20d9-4b6d-b097-d14b9534faa1"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.842217 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/267a25c3-20d9-4b6d-b097-d14b9534faa1-kube-api-access-9mvlv" (OuterVolumeSpecName: "kube-api-access-9mvlv") pod "267a25c3-20d9-4b6d-b097-d14b9534faa1" (UID: "267a25c3-20d9-4b6d-b097-d14b9534faa1"). InnerVolumeSpecName "kube-api-access-9mvlv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.849400 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/267a25c3-20d9-4b6d-b097-d14b9534faa1-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "267a25c3-20d9-4b6d-b097-d14b9534faa1" (UID: "267a25c3-20d9-4b6d-b097-d14b9534faa1"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.849480 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "267a25c3-20d9-4b6d-b097-d14b9534faa1" (UID: "267a25c3-20d9-4b6d-b097-d14b9534faa1"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.877189 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-config" (OuterVolumeSpecName: "config") pod "267a25c3-20d9-4b6d-b097-d14b9534faa1" (UID: "267a25c3-20d9-4b6d-b097-d14b9534faa1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.877610 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-web-config" (OuterVolumeSpecName: "web-config") pod "267a25c3-20d9-4b6d-b097-d14b9534faa1" (UID: "267a25c3-20d9-4b6d-b097-d14b9534faa1"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.890902 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "267a25c3-20d9-4b6d-b097-d14b9534faa1" (UID: "267a25c3-20d9-4b6d-b097-d14b9534faa1"). InnerVolumeSpecName "pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.937842 4930 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/267a25c3-20d9-4b6d-b097-d14b9534faa1-tls-assets\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.937873 4930 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.937887 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9mvlv\" (UniqueName: \"kubernetes.io/projected/267a25c3-20d9-4b6d-b097-d14b9534faa1-kube-api-access-9mvlv\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.937898 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.937908 4930 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/267a25c3-20d9-4b6d-b097-d14b9534faa1-web-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.937916 4930 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/267a25c3-20d9-4b6d-b097-d14b9534faa1-config-out\") on node \"crc\" DevicePath \"\"" Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.937947 4930 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\") on node \"crc\" " Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.963325 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-fghdj-config-jppqn"] Mar 13 09:33:59 crc kubenswrapper[4930]: W0313 09:33:59.969514 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda38f82f1_804d_4abb_a0c3_e70fe84a7c94.slice/crio-43116b0a9c827c973a18d34f1fefa864cad7f689e4d846c7ec1887f9149153a6 WatchSource:0}: Error finding container 43116b0a9c827c973a18d34f1fefa864cad7f689e4d846c7ec1887f9149153a6: Status 404 returned error can't find the container with id 43116b0a9c827c973a18d34f1fefa864cad7f689e4d846c7ec1887f9149153a6 Mar 13 09:33:59 crc kubenswrapper[4930]: I0313 09:33:59.986228 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd600dfb-9194-46a8-b272-0d248c6baf24" path="/var/lib/kubelet/pods/cd600dfb-9194-46a8-b272-0d248c6baf24/volumes" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.004042 4930 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.004412 4930 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd") on node "crc" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.039575 4930 reconciler_common.go:293] "Volume detached for volume \"pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.137610 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556574-dj8l8"] Mar 13 09:34:00 crc kubenswrapper[4930]: E0313 09:34:00.138494 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="init-config-reloader" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.138514 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="init-config-reloader" Mar 13 09:34:00 crc kubenswrapper[4930]: E0313 09:34:00.138532 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="prometheus" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.138538 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="prometheus" Mar 13 09:34:00 crc kubenswrapper[4930]: E0313 09:34:00.138565 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="config-reloader" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.138571 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="config-reloader" Mar 13 09:34:00 crc kubenswrapper[4930]: E0313 09:34:00.138582 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="thanos-sidecar" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.138589 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="thanos-sidecar" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.138778 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="thanos-sidecar" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.138795 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="prometheus" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.138807 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" containerName="config-reloader" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.139521 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556574-dj8l8" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.142552 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.142616 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.142792 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.158704 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556574-dj8l8"] Mar 13 09:34:00 crc kubenswrapper[4930]: W0313 09:34:00.198246 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod97bec685_87be_4c44_8678_189f13ffce7c.slice/crio-083825c462c71eba24b362341b5110a57e69de30e1c7c2a5a481d14eaaaa1622 WatchSource:0}: Error finding container 083825c462c71eba24b362341b5110a57e69de30e1c7c2a5a481d14eaaaa1622: Status 404 returned error can't find the container with id 083825c462c71eba24b362341b5110a57e69de30e1c7c2a5a481d14eaaaa1622 Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.207196 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.242895 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hs5br\" (UniqueName: \"kubernetes.io/projected/a2b7d7fd-2404-409b-a6bb-c000f04536cb-kube-api-access-hs5br\") pod \"auto-csr-approver-29556574-dj8l8\" (UID: \"a2b7d7fd-2404-409b-a6bb-c000f04536cb\") " pod="openshift-infra/auto-csr-approver-29556574-dj8l8" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.303532 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"97bec685-87be-4c44-8678-189f13ffce7c","Type":"ContainerStarted","Data":"083825c462c71eba24b362341b5110a57e69de30e1c7c2a5a481d14eaaaa1622"} Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.309779 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj-config-jppqn" event={"ID":"a38f82f1-804d-4abb-a0c3-e70fe84a7c94","Type":"ContainerStarted","Data":"43116b0a9c827c973a18d34f1fefa864cad7f689e4d846c7ec1887f9149153a6"} Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.317113 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"267a25c3-20d9-4b6d-b097-d14b9534faa1","Type":"ContainerDied","Data":"d17eb4e70e7492b1c006c38a129a9180fafabb7d947347a58c0d5d76b3b15fea"} Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.317173 4930 scope.go:117] "RemoveContainer" containerID="f51adb56f7ae4d763421fb6b107e55503b3d99f2cd6a3fecf2ad91a38012ff2b" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.317347 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.345363 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hs5br\" (UniqueName: \"kubernetes.io/projected/a2b7d7fd-2404-409b-a6bb-c000f04536cb-kube-api-access-hs5br\") pod \"auto-csr-approver-29556574-dj8l8\" (UID: \"a2b7d7fd-2404-409b-a6bb-c000f04536cb\") " pod="openshift-infra/auto-csr-approver-29556574-dj8l8" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.348311 4930 scope.go:117] "RemoveContainer" containerID="d523e85de62adeb9f18cd8d5c405896649ac61c40d0f00713e6a33213b02def4" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.353990 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.369550 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.370590 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hs5br\" (UniqueName: \"kubernetes.io/projected/a2b7d7fd-2404-409b-a6bb-c000f04536cb-kube-api-access-hs5br\") pod \"auto-csr-approver-29556574-dj8l8\" (UID: \"a2b7d7fd-2404-409b-a6bb-c000f04536cb\") " pod="openshift-infra/auto-csr-approver-29556574-dj8l8" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.378884 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.380123 4930 scope.go:117] "RemoveContainer" containerID="9aaa51fb4611379e8c17b3648f7eeeccd410a16f268c69209f0374f1300e0f58" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.384811 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.388915 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-gsfzj" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.389341 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.389562 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-2" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.389699 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.389808 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.389904 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-1" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.390002 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.394095 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.397996 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.420697 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.427004 4930 scope.go:117] "RemoveContainer" containerID="36254e9d39ddea7c93391459ef1c96007498228c864b0042412472d83f0d7b83" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.446915 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/54391e09-d1db-450f-9511-bfeaf8b0d817-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.447137 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/54391e09-d1db-450f-9511-bfeaf8b0d817-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.447242 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.447327 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.447544 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/54391e09-d1db-450f-9511-bfeaf8b0d817-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.447649 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.447784 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcq4x\" (UniqueName: \"kubernetes.io/projected/54391e09-d1db-450f-9511-bfeaf8b0d817-kube-api-access-kcq4x\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.447922 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.448160 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-config\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.448292 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/54391e09-d1db-450f-9511-bfeaf8b0d817-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.448381 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.448527 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.448797 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/54391e09-d1db-450f-9511-bfeaf8b0d817-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.478948 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556574-dj8l8" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.552637 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.552701 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/54391e09-d1db-450f-9511-bfeaf8b0d817-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.552738 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/54391e09-d1db-450f-9511-bfeaf8b0d817-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.552785 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/54391e09-d1db-450f-9511-bfeaf8b0d817-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.552801 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.552822 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.552867 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/54391e09-d1db-450f-9511-bfeaf8b0d817-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.552880 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.552898 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcq4x\" (UniqueName: \"kubernetes.io/projected/54391e09-d1db-450f-9511-bfeaf8b0d817-kube-api-access-kcq4x\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.552922 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.552974 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-config\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.553007 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/54391e09-d1db-450f-9511-bfeaf8b0d817-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.553030 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.557099 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-1\" (UniqueName: \"kubernetes.io/configmap/54391e09-d1db-450f-9511-bfeaf8b0d817-prometheus-metric-storage-rulefiles-1\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.557101 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-2\" (UniqueName: \"kubernetes.io/configmap/54391e09-d1db-450f-9511-bfeaf8b0d817-prometheus-metric-storage-rulefiles-2\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.557696 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/54391e09-d1db-450f-9511-bfeaf8b0d817-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.559498 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.559583 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-config\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.561605 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.562229 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.565112 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.565559 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.565600 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e182641cca8f9c7abeba89f962dbe513ebbaf8cfc329d127147730d16d8b04fe/globalmount\"" pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.566075 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/54391e09-d1db-450f-9511-bfeaf8b0d817-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.568123 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/54391e09-d1db-450f-9511-bfeaf8b0d817-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.569720 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/54391e09-d1db-450f-9511-bfeaf8b0d817-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.574667 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcq4x\" (UniqueName: \"kubernetes.io/projected/54391e09-d1db-450f-9511-bfeaf8b0d817-kube-api-access-kcq4x\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.633824 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8959da5f-6972-4913-aac6-e5a2a19b6afd\") pod \"prometheus-metric-storage-0\" (UID: \"54391e09-d1db-450f-9511-bfeaf8b0d817\") " pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:00 crc kubenswrapper[4930]: I0313 09:34:00.716304 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:01 crc kubenswrapper[4930]: I0313 09:34:01.104994 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556574-dj8l8"] Mar 13 09:34:01 crc kubenswrapper[4930]: I0313 09:34:01.173004 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Mar 13 09:34:01 crc kubenswrapper[4930]: I0313 09:34:01.327180 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556574-dj8l8" event={"ID":"a2b7d7fd-2404-409b-a6bb-c000f04536cb","Type":"ContainerStarted","Data":"dfc87dabb0b8e02e342a87b179c822d90eab1540293ab3f536442b01e80f36e1"} Mar 13 09:34:01 crc kubenswrapper[4930]: I0313 09:34:01.329890 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"54391e09-d1db-450f-9511-bfeaf8b0d817","Type":"ContainerStarted","Data":"194c0f067d3eb2479f6234417f4daab690305253aa4aa64c252bcff5dd621bbc"} Mar 13 09:34:01 crc kubenswrapper[4930]: I0313 09:34:01.384120 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:34:01 crc kubenswrapper[4930]: I0313 09:34:01.390043 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/7b288113-babd-4a27-b397-cbdcbfdc195c-etc-swift\") pod \"swift-storage-0\" (UID: \"7b288113-babd-4a27-b397-cbdcbfdc195c\") " pod="openstack/swift-storage-0" Mar 13 09:34:01 crc kubenswrapper[4930]: I0313 09:34:01.684218 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 13 09:34:01 crc kubenswrapper[4930]: I0313 09:34:01.990793 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="267a25c3-20d9-4b6d-b097-d14b9534faa1" path="/var/lib/kubelet/pods/267a25c3-20d9-4b6d-b097-d14b9534faa1/volumes" Mar 13 09:34:02 crc kubenswrapper[4930]: I0313 09:34:02.339485 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 13 09:34:02 crc kubenswrapper[4930]: I0313 09:34:02.346947 4930 generic.go:334] "Generic (PLEG): container finished" podID="a38f82f1-804d-4abb-a0c3-e70fe84a7c94" containerID="90290252b81588f175401788aba185e7d3a83628e294a972697ba4b38ed16e21" exitCode=0 Mar 13 09:34:02 crc kubenswrapper[4930]: I0313 09:34:02.346988 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj-config-jppqn" event={"ID":"a38f82f1-804d-4abb-a0c3-e70fe84a7c94","Type":"ContainerDied","Data":"90290252b81588f175401788aba185e7d3a83628e294a972697ba4b38ed16e21"} Mar 13 09:34:02 crc kubenswrapper[4930]: I0313 09:34:02.592634 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:34:03 crc kubenswrapper[4930]: I0313 09:34:03.025864 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 13 09:34:03 crc kubenswrapper[4930]: I0313 09:34:03.121968 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="a6623f09-f961-411a-9323-8cbf953c7148" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.136:5671: connect: connection refused" Mar 13 09:34:03 crc kubenswrapper[4930]: I0313 09:34:03.160078 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="2f69fa32-5b72-4f9b-9176-d43ee011605d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.137:5671: connect: connection refused" Mar 13 09:34:03 crc kubenswrapper[4930]: I0313 09:34:03.360225 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"d0bf9a766c1ae1b33a9c2ca26178bca30c80841296de03181a2caea7fb1809ee"} Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.379852 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"54391e09-d1db-450f-9511-bfeaf8b0d817","Type":"ContainerStarted","Data":"94247472a91d31873a2d77855344264e3e56bf394658dcd672c44c7f0bdea380"} Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.382262 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556574-dj8l8" event={"ID":"a2b7d7fd-2404-409b-a6bb-c000f04536cb","Type":"ContainerStarted","Data":"be9f2773447a84d88c2d62d4824149378dce13cfe4f15b91538bba9477a016e6"} Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.416217 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556574-dj8l8" podStartSLOduration=4.161746713 podStartE2EDuration="5.416200571s" podCreationTimestamp="2026-03-13 09:34:00 +0000 UTC" firstStartedPulling="2026-03-13 09:34:01.153093697 +0000 UTC m=+1281.903008374" lastFinishedPulling="2026-03-13 09:34:02.407547565 +0000 UTC m=+1283.157462232" observedRunningTime="2026-03-13 09:34:05.411626661 +0000 UTC m=+1286.161541348" watchObservedRunningTime="2026-03-13 09:34:05.416200571 +0000 UTC m=+1286.166115248" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.546228 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.607711 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-scripts\") pod \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.607797 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-run\") pod \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.607840 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cj6fg\" (UniqueName: \"kubernetes.io/projected/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-kube-api-access-cj6fg\") pod \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.607879 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-run" (OuterVolumeSpecName: "var-run") pod "a38f82f1-804d-4abb-a0c3-e70fe84a7c94" (UID: "a38f82f1-804d-4abb-a0c3-e70fe84a7c94"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.607937 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-additional-scripts\") pod \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.607959 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-log-ovn\") pod \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.607979 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-run-ovn\") pod \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\" (UID: \"a38f82f1-804d-4abb-a0c3-e70fe84a7c94\") " Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.608384 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "a38f82f1-804d-4abb-a0c3-e70fe84a7c94" (UID: "a38f82f1-804d-4abb-a0c3-e70fe84a7c94"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.608425 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "a38f82f1-804d-4abb-a0c3-e70fe84a7c94" (UID: "a38f82f1-804d-4abb-a0c3-e70fe84a7c94"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.608867 4930 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-run\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.608879 4930 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.608889 4930 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.608875 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "a38f82f1-804d-4abb-a0c3-e70fe84a7c94" (UID: "a38f82f1-804d-4abb-a0c3-e70fe84a7c94"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.609046 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-scripts" (OuterVolumeSpecName: "scripts") pod "a38f82f1-804d-4abb-a0c3-e70fe84a7c94" (UID: "a38f82f1-804d-4abb-a0c3-e70fe84a7c94"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.616932 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-kube-api-access-cj6fg" (OuterVolumeSpecName: "kube-api-access-cj6fg") pod "a38f82f1-804d-4abb-a0c3-e70fe84a7c94" (UID: "a38f82f1-804d-4abb-a0c3-e70fe84a7c94"). InnerVolumeSpecName "kube-api-access-cj6fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.710886 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.711411 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cj6fg\" (UniqueName: \"kubernetes.io/projected/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-kube-api-access-cj6fg\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:05 crc kubenswrapper[4930]: I0313 09:34:05.711498 4930 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/a38f82f1-804d-4abb-a0c3-e70fe84a7c94-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:06 crc kubenswrapper[4930]: I0313 09:34:06.392533 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-fghdj-config-jppqn" Mar 13 09:34:06 crc kubenswrapper[4930]: I0313 09:34:06.392524 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-fghdj-config-jppqn" event={"ID":"a38f82f1-804d-4abb-a0c3-e70fe84a7c94","Type":"ContainerDied","Data":"43116b0a9c827c973a18d34f1fefa864cad7f689e4d846c7ec1887f9149153a6"} Mar 13 09:34:06 crc kubenswrapper[4930]: I0313 09:34:06.392651 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="43116b0a9c827c973a18d34f1fefa864cad7f689e4d846c7ec1887f9149153a6" Mar 13 09:34:06 crc kubenswrapper[4930]: I0313 09:34:06.394256 4930 generic.go:334] "Generic (PLEG): container finished" podID="a2b7d7fd-2404-409b-a6bb-c000f04536cb" containerID="be9f2773447a84d88c2d62d4824149378dce13cfe4f15b91538bba9477a016e6" exitCode=0 Mar 13 09:34:06 crc kubenswrapper[4930]: I0313 09:34:06.394361 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556574-dj8l8" event={"ID":"a2b7d7fd-2404-409b-a6bb-c000f04536cb","Type":"ContainerDied","Data":"be9f2773447a84d88c2d62d4824149378dce13cfe4f15b91538bba9477a016e6"} Mar 13 09:34:06 crc kubenswrapper[4930]: I0313 09:34:06.633075 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-fghdj-config-jppqn"] Mar 13 09:34:06 crc kubenswrapper[4930]: I0313 09:34:06.644932 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-fghdj-config-jppqn"] Mar 13 09:34:07 crc kubenswrapper[4930]: I0313 09:34:07.982925 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a38f82f1-804d-4abb-a0c3-e70fe84a7c94" path="/var/lib/kubelet/pods/a38f82f1-804d-4abb-a0c3-e70fe84a7c94/volumes" Mar 13 09:34:11 crc kubenswrapper[4930]: I0313 09:34:11.443517 4930 generic.go:334] "Generic (PLEG): container finished" podID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerID="94247472a91d31873a2d77855344264e3e56bf394658dcd672c44c7f0bdea380" exitCode=0 Mar 13 09:34:11 crc kubenswrapper[4930]: I0313 09:34:11.443601 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"54391e09-d1db-450f-9511-bfeaf8b0d817","Type":"ContainerDied","Data":"94247472a91d31873a2d77855344264e3e56bf394658dcd672c44c7f0bdea380"} Mar 13 09:34:12 crc kubenswrapper[4930]: I0313 09:34:12.308458 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:34:12 crc kubenswrapper[4930]: I0313 09:34:12.308528 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:34:12 crc kubenswrapper[4930]: I0313 09:34:12.851314 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 13 09:34:13 crc kubenswrapper[4930]: I0313 09:34:13.120263 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="a6623f09-f961-411a-9323-8cbf953c7148" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.136:5671: connect: connection refused" Mar 13 09:34:13 crc kubenswrapper[4930]: I0313 09:34:13.161644 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-1" Mar 13 09:34:15 crc kubenswrapper[4930]: E0313 09:34:15.138090 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-glance-api:current-podified" Mar 13 09:34:15 crc kubenswrapper[4930]: E0313 09:34:15.138589 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:glance-db-sync,Image:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/glance/glance.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-95f64,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42415,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42415,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-db-sync-g82c9_openstack(bf10d12f-578b-400e-b480-60a7343bc344): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:34:15 crc kubenswrapper[4930]: E0313 09:34:15.139755 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/glance-db-sync-g82c9" podUID="bf10d12f-578b-400e-b480-60a7343bc344" Mar 13 09:34:15 crc kubenswrapper[4930]: I0313 09:34:15.272826 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556574-dj8l8" Mar 13 09:34:15 crc kubenswrapper[4930]: I0313 09:34:15.342853 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hs5br\" (UniqueName: \"kubernetes.io/projected/a2b7d7fd-2404-409b-a6bb-c000f04536cb-kube-api-access-hs5br\") pod \"a2b7d7fd-2404-409b-a6bb-c000f04536cb\" (UID: \"a2b7d7fd-2404-409b-a6bb-c000f04536cb\") " Mar 13 09:34:15 crc kubenswrapper[4930]: I0313 09:34:15.347273 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b7d7fd-2404-409b-a6bb-c000f04536cb-kube-api-access-hs5br" (OuterVolumeSpecName: "kube-api-access-hs5br") pod "a2b7d7fd-2404-409b-a6bb-c000f04536cb" (UID: "a2b7d7fd-2404-409b-a6bb-c000f04536cb"). InnerVolumeSpecName "kube-api-access-hs5br". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:15 crc kubenswrapper[4930]: I0313 09:34:15.445241 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hs5br\" (UniqueName: \"kubernetes.io/projected/a2b7d7fd-2404-409b-a6bb-c000f04536cb-kube-api-access-hs5br\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:15 crc kubenswrapper[4930]: I0313 09:34:15.483547 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556574-dj8l8" Mar 13 09:34:15 crc kubenswrapper[4930]: I0313 09:34:15.484358 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556574-dj8l8" event={"ID":"a2b7d7fd-2404-409b-a6bb-c000f04536cb","Type":"ContainerDied","Data":"dfc87dabb0b8e02e342a87b179c822d90eab1540293ab3f536442b01e80f36e1"} Mar 13 09:34:15 crc kubenswrapper[4930]: I0313 09:34:15.484392 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfc87dabb0b8e02e342a87b179c822d90eab1540293ab3f536442b01e80f36e1" Mar 13 09:34:15 crc kubenswrapper[4930]: E0313 09:34:15.487110 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"glance-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-glance-api:current-podified\\\"\"" pod="openstack/glance-db-sync-g82c9" podUID="bf10d12f-578b-400e-b480-60a7343bc344" Mar 13 09:34:16 crc kubenswrapper[4930]: I0313 09:34:16.341101 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556568-dvxwm"] Mar 13 09:34:16 crc kubenswrapper[4930]: I0313 09:34:16.354551 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556568-dvxwm"] Mar 13 09:34:16 crc kubenswrapper[4930]: I0313 09:34:16.502600 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"2b8c8dbd4f80fe09882ebd1d7c2a45fcb8355a1f7baa9abc55a17f7270c6d984"} Mar 13 09:34:16 crc kubenswrapper[4930]: I0313 09:34:16.503069 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"26b5109af8b471299d07249964a18d927fc9482e20d0ea959c3507a4404535e2"} Mar 13 09:34:16 crc kubenswrapper[4930]: I0313 09:34:16.503090 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"f7d524984c3e68a057f8237185127bf27d991a2d4fce86e954e68d3ddf7aab6e"} Mar 13 09:34:16 crc kubenswrapper[4930]: I0313 09:34:16.503105 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"fccec744cc6756cbf91e380426e2a6caa6d145993bf2a1aef110f0e4b26842a5"} Mar 13 09:34:16 crc kubenswrapper[4930]: I0313 09:34:16.505064 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"97bec685-87be-4c44-8678-189f13ffce7c","Type":"ContainerStarted","Data":"2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4"} Mar 13 09:34:16 crc kubenswrapper[4930]: I0313 09:34:16.509261 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"54391e09-d1db-450f-9511-bfeaf8b0d817","Type":"ContainerStarted","Data":"6eff41a2dd17dda4fc1c039e2a2403c0a42d01c5a48ac061aed74d6715199b83"} Mar 13 09:34:16 crc kubenswrapper[4930]: I0313 09:34:16.528064 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=2.205010297 podStartE2EDuration="17.52804277s" podCreationTimestamp="2026-03-13 09:33:59 +0000 UTC" firstStartedPulling="2026-03-13 09:34:00.208724908 +0000 UTC m=+1280.958639585" lastFinishedPulling="2026-03-13 09:34:15.531757381 +0000 UTC m=+1296.281672058" observedRunningTime="2026-03-13 09:34:16.520579535 +0000 UTC m=+1297.270494222" watchObservedRunningTime="2026-03-13 09:34:16.52804277 +0000 UTC m=+1297.277957467" Mar 13 09:34:17 crc kubenswrapper[4930]: I0313 09:34:17.988914 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8210f60a-65d8-4fa0-9ad7-c187c3393c5e" path="/var/lib/kubelet/pods/8210f60a-65d8-4fa0-9ad7-c187c3393c5e/volumes" Mar 13 09:34:18 crc kubenswrapper[4930]: I0313 09:34:18.528124 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"f2bac681acc246ada0524a3176b7d44f2f113f695eb1189a799b923cf784628e"} Mar 13 09:34:19 crc kubenswrapper[4930]: I0313 09:34:19.538483 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"54391e09-d1db-450f-9511-bfeaf8b0d817","Type":"ContainerStarted","Data":"10e9c01fe2ed232d8c9aaf6d56375fcd123a4d9754e63db5586e6c631a8cb0f1"} Mar 13 09:34:19 crc kubenswrapper[4930]: I0313 09:34:19.538940 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"54391e09-d1db-450f-9511-bfeaf8b0d817","Type":"ContainerStarted","Data":"839f7aae002ff768901d38cfa146f6ea99e5c8e627d92296b63cbcff38d40b6b"} Mar 13 09:34:19 crc kubenswrapper[4930]: I0313 09:34:19.542216 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"c97fe557cf46cdf2e2966c6f73d7232ebbbc3e3144b2a07fc3da30f1cad3674c"} Mar 13 09:34:19 crc kubenswrapper[4930]: I0313 09:34:19.542258 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"bd1998a140db424b9e81c30fcc1e13d656ba49d3e152a78b1bbab3c3ffc7307b"} Mar 13 09:34:19 crc kubenswrapper[4930]: I0313 09:34:19.542267 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"e3f6ac1a4320f222f219a58a44cfa880f1bc33a09b3ef36817872baa29b7f36a"} Mar 13 09:34:19 crc kubenswrapper[4930]: I0313 09:34:19.564539 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=19.564521305 podStartE2EDuration="19.564521305s" podCreationTimestamp="2026-03-13 09:34:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:34:19.56090061 +0000 UTC m=+1300.310815297" watchObservedRunningTime="2026-03-13 09:34:19.564521305 +0000 UTC m=+1300.314435982" Mar 13 09:34:20 crc kubenswrapper[4930]: I0313 09:34:20.554866 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"2c6c7f5a4736f093f3c4172b06d72b567b0f11866b2bcec09685b9ca7c986ae0"} Mar 13 09:34:20 crc kubenswrapper[4930]: I0313 09:34:20.716348 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:21 crc kubenswrapper[4930]: I0313 09:34:21.571186 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"bdb987970f782fc6ffe318fe93a28e73f7e6e2a78a2e202c9b3a52adab929265"} Mar 13 09:34:21 crc kubenswrapper[4930]: I0313 09:34:21.571593 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"4113c03bf615f083c3e3d37d95502c3f6c893de0a678ae4b5588848b8e8e0869"} Mar 13 09:34:21 crc kubenswrapper[4930]: I0313 09:34:21.571613 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"bc9c6a626130bf6fcf517c4d14001ab92245c0cd31573dc8d510fc33d040a8dc"} Mar 13 09:34:21 crc kubenswrapper[4930]: I0313 09:34:21.571631 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"79c3021f500c6c8d265199b3adba88fd6ae015e17b4b450d04a94b210143feac"} Mar 13 09:34:21 crc kubenswrapper[4930]: I0313 09:34:21.571649 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"ea744c6346223ba3d02598b0cd6a73c9f3caf1e18e5f7d86f7ad9cd6df684de0"} Mar 13 09:34:22 crc kubenswrapper[4930]: I0313 09:34:22.589713 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"7b288113-babd-4a27-b397-cbdcbfdc195c","Type":"ContainerStarted","Data":"a1b8f732f6b4d740d524ed1acac246d6940854cae84609ee86e5653ef750b780"} Mar 13 09:34:22 crc kubenswrapper[4930]: I0313 09:34:22.904237 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=37.152081266 podStartE2EDuration="54.904212136s" podCreationTimestamp="2026-03-13 09:33:28 +0000 UTC" firstStartedPulling="2026-03-13 09:34:02.378540545 +0000 UTC m=+1283.128455222" lastFinishedPulling="2026-03-13 09:34:20.130671405 +0000 UTC m=+1300.880586092" observedRunningTime="2026-03-13 09:34:22.633814648 +0000 UTC m=+1303.383729335" watchObservedRunningTime="2026-03-13 09:34:22.904212136 +0000 UTC m=+1303.654126813" Mar 13 09:34:22 crc kubenswrapper[4930]: I0313 09:34:22.911274 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-w66lq"] Mar 13 09:34:22 crc kubenswrapper[4930]: E0313 09:34:22.911740 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b7d7fd-2404-409b-a6bb-c000f04536cb" containerName="oc" Mar 13 09:34:22 crc kubenswrapper[4930]: I0313 09:34:22.911761 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b7d7fd-2404-409b-a6bb-c000f04536cb" containerName="oc" Mar 13 09:34:22 crc kubenswrapper[4930]: E0313 09:34:22.911781 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a38f82f1-804d-4abb-a0c3-e70fe84a7c94" containerName="ovn-config" Mar 13 09:34:22 crc kubenswrapper[4930]: I0313 09:34:22.911790 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a38f82f1-804d-4abb-a0c3-e70fe84a7c94" containerName="ovn-config" Mar 13 09:34:22 crc kubenswrapper[4930]: I0313 09:34:22.912061 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="a38f82f1-804d-4abb-a0c3-e70fe84a7c94" containerName="ovn-config" Mar 13 09:34:22 crc kubenswrapper[4930]: I0313 09:34:22.912085 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2b7d7fd-2404-409b-a6bb-c000f04536cb" containerName="oc" Mar 13 09:34:22 crc kubenswrapper[4930]: I0313 09:34:22.913735 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:22 crc kubenswrapper[4930]: I0313 09:34:22.917273 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 13 09:34:22 crc kubenswrapper[4930]: I0313 09:34:22.935397 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-w66lq"] Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.072161 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.072581 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.072664 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whtfb\" (UniqueName: \"kubernetes.io/projected/31a3e825-3a78-4284-a9ee-56aaa6c06c83-kube-api-access-whtfb\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.072694 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.072717 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.072776 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-config\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.121593 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-2" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.178689 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.178743 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.178828 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-config\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.178912 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.178975 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.179058 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whtfb\" (UniqueName: \"kubernetes.io/projected/31a3e825-3a78-4284-a9ee-56aaa6c06c83-kube-api-access-whtfb\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.179649 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-ovsdbserver-sb\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.179758 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-dns-svc\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.179909 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-dns-swift-storage-0\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.180129 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-config\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.180414 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-ovsdbserver-nb\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.204693 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whtfb\" (UniqueName: \"kubernetes.io/projected/31a3e825-3a78-4284-a9ee-56aaa6c06c83-kube-api-access-whtfb\") pod \"dnsmasq-dns-5c79d794d7-w66lq\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.247706 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.607527 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-ddz9m"] Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.609240 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ddz9m" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.620675 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-ddz9m"] Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.688521 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-pbbn5"] Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.697015 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pbbn5" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.695625 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86a246ca-6999-4a55-aa0b-b723006d6305-operator-scripts\") pod \"cinder-db-create-ddz9m\" (UID: \"86a246ca-6999-4a55-aa0b-b723006d6305\") " pod="openstack/cinder-db-create-ddz9m" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.697615 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9h7v\" (UniqueName: \"kubernetes.io/projected/86a246ca-6999-4a55-aa0b-b723006d6305-kube-api-access-n9h7v\") pod \"cinder-db-create-ddz9m\" (UID: \"86a246ca-6999-4a55-aa0b-b723006d6305\") " pod="openstack/cinder-db-create-ddz9m" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.725682 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-pbbn5"] Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.804699 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd9tj\" (UniqueName: \"kubernetes.io/projected/95e1aca8-117b-43da-b0ce-3b132c4a6c60-kube-api-access-qd9tj\") pod \"heat-db-create-pbbn5\" (UID: \"95e1aca8-117b-43da-b0ce-3b132c4a6c60\") " pod="openstack/heat-db-create-pbbn5" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.804862 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86a246ca-6999-4a55-aa0b-b723006d6305-operator-scripts\") pod \"cinder-db-create-ddz9m\" (UID: \"86a246ca-6999-4a55-aa0b-b723006d6305\") " pod="openstack/cinder-db-create-ddz9m" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.804894 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9h7v\" (UniqueName: \"kubernetes.io/projected/86a246ca-6999-4a55-aa0b-b723006d6305-kube-api-access-n9h7v\") pod \"cinder-db-create-ddz9m\" (UID: \"86a246ca-6999-4a55-aa0b-b723006d6305\") " pod="openstack/cinder-db-create-ddz9m" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.804959 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95e1aca8-117b-43da-b0ce-3b132c4a6c60-operator-scripts\") pod \"heat-db-create-pbbn5\" (UID: \"95e1aca8-117b-43da-b0ce-3b132c4a6c60\") " pod="openstack/heat-db-create-pbbn5" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.805687 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86a246ca-6999-4a55-aa0b-b723006d6305-operator-scripts\") pod \"cinder-db-create-ddz9m\" (UID: \"86a246ca-6999-4a55-aa0b-b723006d6305\") " pod="openstack/cinder-db-create-ddz9m" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.815433 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-0b34-account-create-update-grx5t"] Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.817208 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-0b34-account-create-update-grx5t" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.824915 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.843101 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-0b34-account-create-update-grx5t"] Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.874415 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9h7v\" (UniqueName: \"kubernetes.io/projected/86a246ca-6999-4a55-aa0b-b723006d6305-kube-api-access-n9h7v\") pod \"cinder-db-create-ddz9m\" (UID: \"86a246ca-6999-4a55-aa0b-b723006d6305\") " pod="openstack/cinder-db-create-ddz9m" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.909203 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68wgk\" (UniqueName: \"kubernetes.io/projected/caf815ba-1554-4c46-bc4d-38b8106cced5-kube-api-access-68wgk\") pod \"heat-0b34-account-create-update-grx5t\" (UID: \"caf815ba-1554-4c46-bc4d-38b8106cced5\") " pod="openstack/heat-0b34-account-create-update-grx5t" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.909301 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95e1aca8-117b-43da-b0ce-3b132c4a6c60-operator-scripts\") pod \"heat-db-create-pbbn5\" (UID: \"95e1aca8-117b-43da-b0ce-3b132c4a6c60\") " pod="openstack/heat-db-create-pbbn5" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.909342 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caf815ba-1554-4c46-bc4d-38b8106cced5-operator-scripts\") pod \"heat-0b34-account-create-update-grx5t\" (UID: \"caf815ba-1554-4c46-bc4d-38b8106cced5\") " pod="openstack/heat-0b34-account-create-update-grx5t" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.909405 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd9tj\" (UniqueName: \"kubernetes.io/projected/95e1aca8-117b-43da-b0ce-3b132c4a6c60-kube-api-access-qd9tj\") pod \"heat-db-create-pbbn5\" (UID: \"95e1aca8-117b-43da-b0ce-3b132c4a6c60\") " pod="openstack/heat-db-create-pbbn5" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.910299 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95e1aca8-117b-43da-b0ce-3b132c4a6c60-operator-scripts\") pod \"heat-db-create-pbbn5\" (UID: \"95e1aca8-117b-43da-b0ce-3b132c4a6c60\") " pod="openstack/heat-db-create-pbbn5" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.945183 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-d7cf-account-create-update-v5gw9"] Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.946572 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d7cf-account-create-update-v5gw9" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.947886 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ddz9m" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.957283 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-w66lq"] Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.966841 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d7cf-account-create-update-v5gw9"] Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.976898 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd9tj\" (UniqueName: \"kubernetes.io/projected/95e1aca8-117b-43da-b0ce-3b132c4a6c60-kube-api-access-qd9tj\") pod \"heat-db-create-pbbn5\" (UID: \"95e1aca8-117b-43da-b0ce-3b132c4a6c60\") " pod="openstack/heat-db-create-pbbn5" Mar 13 09:34:23 crc kubenswrapper[4930]: I0313 09:34:23.977270 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.013976 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caf815ba-1554-4c46-bc4d-38b8106cced5-operator-scripts\") pod \"heat-0b34-account-create-update-grx5t\" (UID: \"caf815ba-1554-4c46-bc4d-38b8106cced5\") " pod="openstack/heat-0b34-account-create-update-grx5t" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.014305 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f31cf24-9bb2-41c6-a825-c5481a27ffb0-operator-scripts\") pod \"cinder-d7cf-account-create-update-v5gw9\" (UID: \"8f31cf24-9bb2-41c6-a825-c5481a27ffb0\") " pod="openstack/cinder-d7cf-account-create-update-v5gw9" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.014490 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csqn5\" (UniqueName: \"kubernetes.io/projected/8f31cf24-9bb2-41c6-a825-c5481a27ffb0-kube-api-access-csqn5\") pod \"cinder-d7cf-account-create-update-v5gw9\" (UID: \"8f31cf24-9bb2-41c6-a825-c5481a27ffb0\") " pod="openstack/cinder-d7cf-account-create-update-v5gw9" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.014711 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68wgk\" (UniqueName: \"kubernetes.io/projected/caf815ba-1554-4c46-bc4d-38b8106cced5-kube-api-access-68wgk\") pod \"heat-0b34-account-create-update-grx5t\" (UID: \"caf815ba-1554-4c46-bc4d-38b8106cced5\") " pod="openstack/heat-0b34-account-create-update-grx5t" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.016244 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caf815ba-1554-4c46-bc4d-38b8106cced5-operator-scripts\") pod \"heat-0b34-account-create-update-grx5t\" (UID: \"caf815ba-1554-4c46-bc4d-38b8106cced5\") " pod="openstack/heat-0b34-account-create-update-grx5t" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.032770 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pbbn5" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.043853 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68wgk\" (UniqueName: \"kubernetes.io/projected/caf815ba-1554-4c46-bc4d-38b8106cced5-kube-api-access-68wgk\") pod \"heat-0b34-account-create-update-grx5t\" (UID: \"caf815ba-1554-4c46-bc4d-38b8106cced5\") " pod="openstack/heat-0b34-account-create-update-grx5t" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.085562 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-lgr8r"] Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.095424 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lgr8r" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.107731 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lgr8r"] Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.116935 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csqn5\" (UniqueName: \"kubernetes.io/projected/8f31cf24-9bb2-41c6-a825-c5481a27ffb0-kube-api-access-csqn5\") pod \"cinder-d7cf-account-create-update-v5gw9\" (UID: \"8f31cf24-9bb2-41c6-a825-c5481a27ffb0\") " pod="openstack/cinder-d7cf-account-create-update-v5gw9" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.117163 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f31cf24-9bb2-41c6-a825-c5481a27ffb0-operator-scripts\") pod \"cinder-d7cf-account-create-update-v5gw9\" (UID: \"8f31cf24-9bb2-41c6-a825-c5481a27ffb0\") " pod="openstack/cinder-d7cf-account-create-update-v5gw9" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.117952 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f31cf24-9bb2-41c6-a825-c5481a27ffb0-operator-scripts\") pod \"cinder-d7cf-account-create-update-v5gw9\" (UID: \"8f31cf24-9bb2-41c6-a825-c5481a27ffb0\") " pod="openstack/cinder-d7cf-account-create-update-v5gw9" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.139229 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csqn5\" (UniqueName: \"kubernetes.io/projected/8f31cf24-9bb2-41c6-a825-c5481a27ffb0-kube-api-access-csqn5\") pod \"cinder-d7cf-account-create-update-v5gw9\" (UID: \"8f31cf24-9bb2-41c6-a825-c5481a27ffb0\") " pod="openstack/cinder-d7cf-account-create-update-v5gw9" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.143660 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-0b34-account-create-update-grx5t" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.206705 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-vr4gs"] Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.214761 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.218333 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.218710 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fffd6b7f-d18d-4ff8-a446-032de1c3e288-operator-scripts\") pod \"barbican-db-create-lgr8r\" (UID: \"fffd6b7f-d18d-4ff8-a446-032de1c3e288\") " pod="openstack/barbican-db-create-lgr8r" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.218764 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.218785 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ws54h" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.218799 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjzzf\" (UniqueName: \"kubernetes.io/projected/fffd6b7f-d18d-4ff8-a446-032de1c3e288-kube-api-access-kjzzf\") pod \"barbican-db-create-lgr8r\" (UID: \"fffd6b7f-d18d-4ff8-a446-032de1c3e288\") " pod="openstack/barbican-db-create-lgr8r" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.218908 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.245662 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-vr4gs"] Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.300452 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-bdfd-account-create-update-9lgz8"] Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.302115 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bdfd-account-create-update-9lgz8" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.304360 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.318566 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d7cf-account-create-update-v5gw9" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.320118 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkpzh\" (UniqueName: \"kubernetes.io/projected/d0127857-fa77-4179-ad95-7dfb2585169a-kube-api-access-dkpzh\") pod \"keystone-db-sync-vr4gs\" (UID: \"d0127857-fa77-4179-ad95-7dfb2585169a\") " pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.323321 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fffd6b7f-d18d-4ff8-a446-032de1c3e288-operator-scripts\") pod \"barbican-db-create-lgr8r\" (UID: \"fffd6b7f-d18d-4ff8-a446-032de1c3e288\") " pod="openstack/barbican-db-create-lgr8r" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.323523 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0127857-fa77-4179-ad95-7dfb2585169a-combined-ca-bundle\") pod \"keystone-db-sync-vr4gs\" (UID: \"d0127857-fa77-4179-ad95-7dfb2585169a\") " pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.323613 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0127857-fa77-4179-ad95-7dfb2585169a-config-data\") pod \"keystone-db-sync-vr4gs\" (UID: \"d0127857-fa77-4179-ad95-7dfb2585169a\") " pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.324782 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjzzf\" (UniqueName: \"kubernetes.io/projected/fffd6b7f-d18d-4ff8-a446-032de1c3e288-kube-api-access-kjzzf\") pod \"barbican-db-create-lgr8r\" (UID: \"fffd6b7f-d18d-4ff8-a446-032de1c3e288\") " pod="openstack/barbican-db-create-lgr8r" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.326177 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fffd6b7f-d18d-4ff8-a446-032de1c3e288-operator-scripts\") pod \"barbican-db-create-lgr8r\" (UID: \"fffd6b7f-d18d-4ff8-a446-032de1c3e288\") " pod="openstack/barbican-db-create-lgr8r" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.362634 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-zqbhp"] Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.364336 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zqbhp" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.366382 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjzzf\" (UniqueName: \"kubernetes.io/projected/fffd6b7f-d18d-4ff8-a446-032de1c3e288-kube-api-access-kjzzf\") pod \"barbican-db-create-lgr8r\" (UID: \"fffd6b7f-d18d-4ff8-a446-032de1c3e288\") " pod="openstack/barbican-db-create-lgr8r" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.410383 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-bdfd-account-create-update-9lgz8"] Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.422281 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lgr8r" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.428047 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlgrm\" (UniqueName: \"kubernetes.io/projected/f84311c2-0fb2-4d15-8272-f245fcbc9957-kube-api-access-tlgrm\") pod \"neutron-db-create-zqbhp\" (UID: \"f84311c2-0fb2-4d15-8272-f245fcbc9957\") " pod="openstack/neutron-db-create-zqbhp" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.430653 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz8wn\" (UniqueName: \"kubernetes.io/projected/ade20871-d4bc-4a8f-9635-d28bbae44421-kube-api-access-hz8wn\") pod \"barbican-bdfd-account-create-update-9lgz8\" (UID: \"ade20871-d4bc-4a8f-9635-d28bbae44421\") " pod="openstack/barbican-bdfd-account-create-update-9lgz8" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.430846 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f84311c2-0fb2-4d15-8272-f245fcbc9957-operator-scripts\") pod \"neutron-db-create-zqbhp\" (UID: \"f84311c2-0fb2-4d15-8272-f245fcbc9957\") " pod="openstack/neutron-db-create-zqbhp" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.430986 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkpzh\" (UniqueName: \"kubernetes.io/projected/d0127857-fa77-4179-ad95-7dfb2585169a-kube-api-access-dkpzh\") pod \"keystone-db-sync-vr4gs\" (UID: \"d0127857-fa77-4179-ad95-7dfb2585169a\") " pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.434409 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0127857-fa77-4179-ad95-7dfb2585169a-combined-ca-bundle\") pod \"keystone-db-sync-vr4gs\" (UID: \"d0127857-fa77-4179-ad95-7dfb2585169a\") " pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.434538 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0127857-fa77-4179-ad95-7dfb2585169a-config-data\") pod \"keystone-db-sync-vr4gs\" (UID: \"d0127857-fa77-4179-ad95-7dfb2585169a\") " pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.434798 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ade20871-d4bc-4a8f-9635-d28bbae44421-operator-scripts\") pod \"barbican-bdfd-account-create-update-9lgz8\" (UID: \"ade20871-d4bc-4a8f-9635-d28bbae44421\") " pod="openstack/barbican-bdfd-account-create-update-9lgz8" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.449561 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0127857-fa77-4179-ad95-7dfb2585169a-combined-ca-bundle\") pod \"keystone-db-sync-vr4gs\" (UID: \"d0127857-fa77-4179-ad95-7dfb2585169a\") " pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.473298 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0127857-fa77-4179-ad95-7dfb2585169a-config-data\") pod \"keystone-db-sync-vr4gs\" (UID: \"d0127857-fa77-4179-ad95-7dfb2585169a\") " pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.481077 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkpzh\" (UniqueName: \"kubernetes.io/projected/d0127857-fa77-4179-ad95-7dfb2585169a-kube-api-access-dkpzh\") pod \"keystone-db-sync-vr4gs\" (UID: \"d0127857-fa77-4179-ad95-7dfb2585169a\") " pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.511338 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zqbhp"] Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.539644 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlgrm\" (UniqueName: \"kubernetes.io/projected/f84311c2-0fb2-4d15-8272-f245fcbc9957-kube-api-access-tlgrm\") pod \"neutron-db-create-zqbhp\" (UID: \"f84311c2-0fb2-4d15-8272-f245fcbc9957\") " pod="openstack/neutron-db-create-zqbhp" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.539714 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz8wn\" (UniqueName: \"kubernetes.io/projected/ade20871-d4bc-4a8f-9635-d28bbae44421-kube-api-access-hz8wn\") pod \"barbican-bdfd-account-create-update-9lgz8\" (UID: \"ade20871-d4bc-4a8f-9635-d28bbae44421\") " pod="openstack/barbican-bdfd-account-create-update-9lgz8" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.539740 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f84311c2-0fb2-4d15-8272-f245fcbc9957-operator-scripts\") pod \"neutron-db-create-zqbhp\" (UID: \"f84311c2-0fb2-4d15-8272-f245fcbc9957\") " pod="openstack/neutron-db-create-zqbhp" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.539860 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ade20871-d4bc-4a8f-9635-d28bbae44421-operator-scripts\") pod \"barbican-bdfd-account-create-update-9lgz8\" (UID: \"ade20871-d4bc-4a8f-9635-d28bbae44421\") " pod="openstack/barbican-bdfd-account-create-update-9lgz8" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.541321 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ade20871-d4bc-4a8f-9635-d28bbae44421-operator-scripts\") pod \"barbican-bdfd-account-create-update-9lgz8\" (UID: \"ade20871-d4bc-4a8f-9635-d28bbae44421\") " pod="openstack/barbican-bdfd-account-create-update-9lgz8" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.541868 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f84311c2-0fb2-4d15-8272-f245fcbc9957-operator-scripts\") pod \"neutron-db-create-zqbhp\" (UID: \"f84311c2-0fb2-4d15-8272-f245fcbc9957\") " pod="openstack/neutron-db-create-zqbhp" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.554788 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-728c-account-create-update-q5spq"] Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.556430 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-728c-account-create-update-q5spq" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.562812 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.576434 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlgrm\" (UniqueName: \"kubernetes.io/projected/f84311c2-0fb2-4d15-8272-f245fcbc9957-kube-api-access-tlgrm\") pod \"neutron-db-create-zqbhp\" (UID: \"f84311c2-0fb2-4d15-8272-f245fcbc9957\") " pod="openstack/neutron-db-create-zqbhp" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.637289 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" event={"ID":"31a3e825-3a78-4284-a9ee-56aaa6c06c83","Type":"ContainerStarted","Data":"0a5ff13ada066aa843658cef3bc8aaa5d218ebf5d3ce0d7f549fc0e4e90a5096"} Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.645056 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6164de25-5015-4f67-8b21-d0203eae1351-operator-scripts\") pod \"neutron-728c-account-create-update-q5spq\" (UID: \"6164de25-5015-4f67-8b21-d0203eae1351\") " pod="openstack/neutron-728c-account-create-update-q5spq" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.645313 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mql5w\" (UniqueName: \"kubernetes.io/projected/6164de25-5015-4f67-8b21-d0203eae1351-kube-api-access-mql5w\") pod \"neutron-728c-account-create-update-q5spq\" (UID: \"6164de25-5015-4f67-8b21-d0203eae1351\") " pod="openstack/neutron-728c-account-create-update-q5spq" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.645802 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.657531 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-728c-account-create-update-q5spq"] Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.663308 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz8wn\" (UniqueName: \"kubernetes.io/projected/ade20871-d4bc-4a8f-9635-d28bbae44421-kube-api-access-hz8wn\") pod \"barbican-bdfd-account-create-update-9lgz8\" (UID: \"ade20871-d4bc-4a8f-9635-d28bbae44421\") " pod="openstack/barbican-bdfd-account-create-update-9lgz8" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.747614 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mql5w\" (UniqueName: \"kubernetes.io/projected/6164de25-5015-4f67-8b21-d0203eae1351-kube-api-access-mql5w\") pod \"neutron-728c-account-create-update-q5spq\" (UID: \"6164de25-5015-4f67-8b21-d0203eae1351\") " pod="openstack/neutron-728c-account-create-update-q5spq" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.747977 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6164de25-5015-4f67-8b21-d0203eae1351-operator-scripts\") pod \"neutron-728c-account-create-update-q5spq\" (UID: \"6164de25-5015-4f67-8b21-d0203eae1351\") " pod="openstack/neutron-728c-account-create-update-q5spq" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.748743 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6164de25-5015-4f67-8b21-d0203eae1351-operator-scripts\") pod \"neutron-728c-account-create-update-q5spq\" (UID: \"6164de25-5015-4f67-8b21-d0203eae1351\") " pod="openstack/neutron-728c-account-create-update-q5spq" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.786015 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mql5w\" (UniqueName: \"kubernetes.io/projected/6164de25-5015-4f67-8b21-d0203eae1351-kube-api-access-mql5w\") pod \"neutron-728c-account-create-update-q5spq\" (UID: \"6164de25-5015-4f67-8b21-d0203eae1351\") " pod="openstack/neutron-728c-account-create-update-q5spq" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.812520 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bdfd-account-create-update-9lgz8" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.821744 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-ddz9m"] Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.832356 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zqbhp" Mar 13 09:34:24 crc kubenswrapper[4930]: I0313 09:34:24.988405 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-728c-account-create-update-q5spq" Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.227528 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-pbbn5"] Mar 13 09:34:25 crc kubenswrapper[4930]: W0313 09:34:25.237378 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod95e1aca8_117b_43da_b0ce_3b132c4a6c60.slice/crio-5541690f3f90095c1ef96b23733bf3c7c0eae169eacc7e07f0762b2abd83f741 WatchSource:0}: Error finding container 5541690f3f90095c1ef96b23733bf3c7c0eae169eacc7e07f0762b2abd83f741: Status 404 returned error can't find the container with id 5541690f3f90095c1ef96b23733bf3c7c0eae169eacc7e07f0762b2abd83f741 Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.248051 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-d7cf-account-create-update-v5gw9"] Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.453252 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-0b34-account-create-update-grx5t"] Mar 13 09:34:25 crc kubenswrapper[4930]: W0313 09:34:25.464537 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcaf815ba_1554_4c46_bc4d_38b8106cced5.slice/crio-f8ea5ffcdc876b8a09c5e0549151c76fae75855ddb818a1dd2ad1ffbeb406e04 WatchSource:0}: Error finding container f8ea5ffcdc876b8a09c5e0549151c76fae75855ddb818a1dd2ad1ffbeb406e04: Status 404 returned error can't find the container with id f8ea5ffcdc876b8a09c5e0549151c76fae75855ddb818a1dd2ad1ffbeb406e04 Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.591104 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-vr4gs"] Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.635630 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-lgr8r"] Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.667057 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d7cf-account-create-update-v5gw9" event={"ID":"8f31cf24-9bb2-41c6-a825-c5481a27ffb0","Type":"ContainerStarted","Data":"782485b9d9f78d2c982187e6bb785310dfbbf0b35c611d44d7585c989975a169"} Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.667093 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d7cf-account-create-update-v5gw9" event={"ID":"8f31cf24-9bb2-41c6-a825-c5481a27ffb0","Type":"ContainerStarted","Data":"d85325bcd8670a5a5e670698aec9beb3c0f4e8f5b6fe4afe651f2d18e8d4bba0"} Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.672689 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vr4gs" event={"ID":"d0127857-fa77-4179-ad95-7dfb2585169a","Type":"ContainerStarted","Data":"0a4bfca440c9a81c9dbe77e508b4139614be177743a1d5bd23f0a45ed621cca8"} Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.677251 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-0b34-account-create-update-grx5t" event={"ID":"caf815ba-1554-4c46-bc4d-38b8106cced5","Type":"ContainerStarted","Data":"f8ea5ffcdc876b8a09c5e0549151c76fae75855ddb818a1dd2ad1ffbeb406e04"} Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.688457 4930 generic.go:334] "Generic (PLEG): container finished" podID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerID="5460f65f3e2b8821d1a4bdb4576a88d3abd7232ee3465848f2d08959151b5b0d" exitCode=0 Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.688642 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" event={"ID":"31a3e825-3a78-4284-a9ee-56aaa6c06c83","Type":"ContainerDied","Data":"5460f65f3e2b8821d1a4bdb4576a88d3abd7232ee3465848f2d08959151b5b0d"} Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.693784 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-d7cf-account-create-update-v5gw9" podStartSLOduration=2.693762927 podStartE2EDuration="2.693762927s" podCreationTimestamp="2026-03-13 09:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:34:25.678846286 +0000 UTC m=+1306.428760953" watchObservedRunningTime="2026-03-13 09:34:25.693762927 +0000 UTC m=+1306.443677604" Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.700723 4930 generic.go:334] "Generic (PLEG): container finished" podID="86a246ca-6999-4a55-aa0b-b723006d6305" containerID="7e38480a43a6e1adcd53ab5a8cfc002475af7a65aca6c38215a2af9876233bc2" exitCode=0 Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.700792 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ddz9m" event={"ID":"86a246ca-6999-4a55-aa0b-b723006d6305","Type":"ContainerDied","Data":"7e38480a43a6e1adcd53ab5a8cfc002475af7a65aca6c38215a2af9876233bc2"} Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.700817 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ddz9m" event={"ID":"86a246ca-6999-4a55-aa0b-b723006d6305","Type":"ContainerStarted","Data":"27c801f89ff38556e684efff10b9c9b3852e2c49eab7d144c3550cf05a50e257"} Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.702662 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pbbn5" event={"ID":"95e1aca8-117b-43da-b0ce-3b132c4a6c60","Type":"ContainerStarted","Data":"afa291e400ac690b8ecca6419a12980d873fe93d54b442559582f30cd0bf61fb"} Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.702697 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pbbn5" event={"ID":"95e1aca8-117b-43da-b0ce-3b132c4a6c60","Type":"ContainerStarted","Data":"5541690f3f90095c1ef96b23733bf3c7c0eae169eacc7e07f0762b2abd83f741"} Mar 13 09:34:25 crc kubenswrapper[4930]: I0313 09:34:25.756044 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-pbbn5" podStartSLOduration=2.756024207 podStartE2EDuration="2.756024207s" podCreationTimestamp="2026-03-13 09:34:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:34:25.731569116 +0000 UTC m=+1306.481483803" watchObservedRunningTime="2026-03-13 09:34:25.756024207 +0000 UTC m=+1306.505938884" Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.015935 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zqbhp"] Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.030014 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-bdfd-account-create-update-9lgz8"] Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.039944 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-728c-account-create-update-q5spq"] Mar 13 09:34:26 crc kubenswrapper[4930]: W0313 09:34:26.060215 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6164de25_5015_4f67_8b21_d0203eae1351.slice/crio-420033fd92e9fa763c3761ecdd60c58d9a8f0110bdf9ffa53eb04dc6561cbff0 WatchSource:0}: Error finding container 420033fd92e9fa763c3761ecdd60c58d9a8f0110bdf9ffa53eb04dc6561cbff0: Status 404 returned error can't find the container with id 420033fd92e9fa763c3761ecdd60c58d9a8f0110bdf9ffa53eb04dc6561cbff0 Mar 13 09:34:26 crc kubenswrapper[4930]: W0313 09:34:26.060628 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podade20871_d4bc_4a8f_9635_d28bbae44421.slice/crio-5705f3fea3e0c7373e10a0ea8cd33e6463ef59ce1e5b41d4355ffb5cc5635fd7 WatchSource:0}: Error finding container 5705f3fea3e0c7373e10a0ea8cd33e6463ef59ce1e5b41d4355ffb5cc5635fd7: Status 404 returned error can't find the container with id 5705f3fea3e0c7373e10a0ea8cd33e6463ef59ce1e5b41d4355ffb5cc5635fd7 Mar 13 09:34:26 crc kubenswrapper[4930]: W0313 09:34:26.060977 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf84311c2_0fb2_4d15_8272_f245fcbc9957.slice/crio-f0ba084bbb6777a6f08f4f6a9b73305a118dd6797a2e7d80c6e1399e46f2e08e WatchSource:0}: Error finding container f0ba084bbb6777a6f08f4f6a9b73305a118dd6797a2e7d80c6e1399e46f2e08e: Status 404 returned error can't find the container with id f0ba084bbb6777a6f08f4f6a9b73305a118dd6797a2e7d80c6e1399e46f2e08e Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.724287 4930 generic.go:334] "Generic (PLEG): container finished" podID="95e1aca8-117b-43da-b0ce-3b132c4a6c60" containerID="afa291e400ac690b8ecca6419a12980d873fe93d54b442559582f30cd0bf61fb" exitCode=0 Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.724749 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pbbn5" event={"ID":"95e1aca8-117b-43da-b0ce-3b132c4a6c60","Type":"ContainerDied","Data":"afa291e400ac690b8ecca6419a12980d873fe93d54b442559582f30cd0bf61fb"} Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.726739 4930 generic.go:334] "Generic (PLEG): container finished" podID="6164de25-5015-4f67-8b21-d0203eae1351" containerID="2d261f4b77a94477c1f9a893bd96a0a2e94a766d584dc966db7302f165125734" exitCode=0 Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.726792 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-728c-account-create-update-q5spq" event={"ID":"6164de25-5015-4f67-8b21-d0203eae1351","Type":"ContainerDied","Data":"2d261f4b77a94477c1f9a893bd96a0a2e94a766d584dc966db7302f165125734"} Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.726808 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-728c-account-create-update-q5spq" event={"ID":"6164de25-5015-4f67-8b21-d0203eae1351","Type":"ContainerStarted","Data":"420033fd92e9fa763c3761ecdd60c58d9a8f0110bdf9ffa53eb04dc6561cbff0"} Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.729242 4930 generic.go:334] "Generic (PLEG): container finished" podID="8f31cf24-9bb2-41c6-a825-c5481a27ffb0" containerID="782485b9d9f78d2c982187e6bb785310dfbbf0b35c611d44d7585c989975a169" exitCode=0 Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.729333 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d7cf-account-create-update-v5gw9" event={"ID":"8f31cf24-9bb2-41c6-a825-c5481a27ffb0","Type":"ContainerDied","Data":"782485b9d9f78d2c982187e6bb785310dfbbf0b35c611d44d7585c989975a169"} Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.731638 4930 generic.go:334] "Generic (PLEG): container finished" podID="ade20871-d4bc-4a8f-9635-d28bbae44421" containerID="189d6226a00c53bd0ea70ed4d33d2fa9affbfec3cb10f9c48b7b12eb5fe384d6" exitCode=0 Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.731727 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bdfd-account-create-update-9lgz8" event={"ID":"ade20871-d4bc-4a8f-9635-d28bbae44421","Type":"ContainerDied","Data":"189d6226a00c53bd0ea70ed4d33d2fa9affbfec3cb10f9c48b7b12eb5fe384d6"} Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.731880 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bdfd-account-create-update-9lgz8" event={"ID":"ade20871-d4bc-4a8f-9635-d28bbae44421","Type":"ContainerStarted","Data":"5705f3fea3e0c7373e10a0ea8cd33e6463ef59ce1e5b41d4355ffb5cc5635fd7"} Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.737918 4930 generic.go:334] "Generic (PLEG): container finished" podID="caf815ba-1554-4c46-bc4d-38b8106cced5" containerID="2d49098b9512d6183bdf64a950a657e8445ed19536c9fc23675722560104044b" exitCode=0 Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.738008 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-0b34-account-create-update-grx5t" event={"ID":"caf815ba-1554-4c46-bc4d-38b8106cced5","Type":"ContainerDied","Data":"2d49098b9512d6183bdf64a950a657e8445ed19536c9fc23675722560104044b"} Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.743550 4930 generic.go:334] "Generic (PLEG): container finished" podID="fffd6b7f-d18d-4ff8-a446-032de1c3e288" containerID="21eed8e0619bb514a23b49eb0bab1cf671d8a7d01d4d46d3b1b3946831c8583f" exitCode=0 Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.743734 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lgr8r" event={"ID":"fffd6b7f-d18d-4ff8-a446-032de1c3e288","Type":"ContainerDied","Data":"21eed8e0619bb514a23b49eb0bab1cf671d8a7d01d4d46d3b1b3946831c8583f"} Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.743763 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lgr8r" event={"ID":"fffd6b7f-d18d-4ff8-a446-032de1c3e288","Type":"ContainerStarted","Data":"1d6f40093a1a3c19311515e1fff793c70eb3a2dfa93434ab89887f6bd4925f55"} Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.758251 4930 generic.go:334] "Generic (PLEG): container finished" podID="f84311c2-0fb2-4d15-8272-f245fcbc9957" containerID="dd293d0f1b807793ae505f3f997c54e9ecb690d62a8ec6da2b542b321eb582ab" exitCode=0 Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.758329 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zqbhp" event={"ID":"f84311c2-0fb2-4d15-8272-f245fcbc9957","Type":"ContainerDied","Data":"dd293d0f1b807793ae505f3f997c54e9ecb690d62a8ec6da2b542b321eb582ab"} Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.758396 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zqbhp" event={"ID":"f84311c2-0fb2-4d15-8272-f245fcbc9957","Type":"ContainerStarted","Data":"f0ba084bbb6777a6f08f4f6a9b73305a118dd6797a2e7d80c6e1399e46f2e08e"} Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.765701 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" event={"ID":"31a3e825-3a78-4284-a9ee-56aaa6c06c83","Type":"ContainerStarted","Data":"224cc21ace1e5ad0ff39b7f7a7b1da206852e0525c50eaa2f75e871eea068c02"} Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.765771 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:26 crc kubenswrapper[4930]: I0313 09:34:26.870790 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" podStartSLOduration=4.870767537 podStartE2EDuration="4.870767537s" podCreationTimestamp="2026-03-13 09:34:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:34:26.828194512 +0000 UTC m=+1307.578109189" watchObservedRunningTime="2026-03-13 09:34:26.870767537 +0000 UTC m=+1307.620682214" Mar 13 09:34:27 crc kubenswrapper[4930]: I0313 09:34:27.310276 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ddz9m" Mar 13 09:34:27 crc kubenswrapper[4930]: I0313 09:34:27.409661 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9h7v\" (UniqueName: \"kubernetes.io/projected/86a246ca-6999-4a55-aa0b-b723006d6305-kube-api-access-n9h7v\") pod \"86a246ca-6999-4a55-aa0b-b723006d6305\" (UID: \"86a246ca-6999-4a55-aa0b-b723006d6305\") " Mar 13 09:34:27 crc kubenswrapper[4930]: I0313 09:34:27.410264 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86a246ca-6999-4a55-aa0b-b723006d6305-operator-scripts\") pod \"86a246ca-6999-4a55-aa0b-b723006d6305\" (UID: \"86a246ca-6999-4a55-aa0b-b723006d6305\") " Mar 13 09:34:27 crc kubenswrapper[4930]: I0313 09:34:27.411346 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/86a246ca-6999-4a55-aa0b-b723006d6305-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "86a246ca-6999-4a55-aa0b-b723006d6305" (UID: "86a246ca-6999-4a55-aa0b-b723006d6305"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:27 crc kubenswrapper[4930]: I0313 09:34:27.418965 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86a246ca-6999-4a55-aa0b-b723006d6305-kube-api-access-n9h7v" (OuterVolumeSpecName: "kube-api-access-n9h7v") pod "86a246ca-6999-4a55-aa0b-b723006d6305" (UID: "86a246ca-6999-4a55-aa0b-b723006d6305"). InnerVolumeSpecName "kube-api-access-n9h7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:27 crc kubenswrapper[4930]: I0313 09:34:27.512266 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9h7v\" (UniqueName: \"kubernetes.io/projected/86a246ca-6999-4a55-aa0b-b723006d6305-kube-api-access-n9h7v\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:27 crc kubenswrapper[4930]: I0313 09:34:27.512323 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/86a246ca-6999-4a55-aa0b-b723006d6305-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:27 crc kubenswrapper[4930]: I0313 09:34:27.801863 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-ddz9m" Mar 13 09:34:27 crc kubenswrapper[4930]: I0313 09:34:27.805654 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-ddz9m" event={"ID":"86a246ca-6999-4a55-aa0b-b723006d6305","Type":"ContainerDied","Data":"27c801f89ff38556e684efff10b9c9b3852e2c49eab7d144c3550cf05a50e257"} Mar 13 09:34:27 crc kubenswrapper[4930]: I0313 09:34:27.805685 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="27c801f89ff38556e684efff10b9c9b3852e2c49eab7d144c3550cf05a50e257" Mar 13 09:34:28 crc kubenswrapper[4930]: I0313 09:34:28.814783 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g82c9" event={"ID":"bf10d12f-578b-400e-b480-60a7343bc344","Type":"ContainerStarted","Data":"9cd8e9f98e28ab2aa6c6fa08257cd36d1c89eaf1805033449f57940002c31365"} Mar 13 09:34:28 crc kubenswrapper[4930]: I0313 09:34:28.842077 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-g82c9" podStartSLOduration=2.842558974 podStartE2EDuration="37.842057129s" podCreationTimestamp="2026-03-13 09:33:51 +0000 UTC" firstStartedPulling="2026-03-13 09:33:52.402485887 +0000 UTC m=+1273.152400564" lastFinishedPulling="2026-03-13 09:34:27.401984032 +0000 UTC m=+1308.151898719" observedRunningTime="2026-03-13 09:34:28.831822241 +0000 UTC m=+1309.581736928" watchObservedRunningTime="2026-03-13 09:34:28.842057129 +0000 UTC m=+1309.591971806" Mar 13 09:34:30 crc kubenswrapper[4930]: I0313 09:34:30.717668 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:30 crc kubenswrapper[4930]: I0313 09:34:30.723914 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:30 crc kubenswrapper[4930]: I0313 09:34:30.837811 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.242568 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bdfd-account-create-update-9lgz8" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.248639 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zqbhp" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.256315 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-728c-account-create-update-q5spq" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.276094 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lgr8r" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.314986 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-0b34-account-create-update-grx5t" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.323933 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pbbn5" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.325332 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tlgrm\" (UniqueName: \"kubernetes.io/projected/f84311c2-0fb2-4d15-8272-f245fcbc9957-kube-api-access-tlgrm\") pod \"f84311c2-0fb2-4d15-8272-f245fcbc9957\" (UID: \"f84311c2-0fb2-4d15-8272-f245fcbc9957\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.325406 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hz8wn\" (UniqueName: \"kubernetes.io/projected/ade20871-d4bc-4a8f-9635-d28bbae44421-kube-api-access-hz8wn\") pod \"ade20871-d4bc-4a8f-9635-d28bbae44421\" (UID: \"ade20871-d4bc-4a8f-9635-d28bbae44421\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.325482 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6164de25-5015-4f67-8b21-d0203eae1351-operator-scripts\") pod \"6164de25-5015-4f67-8b21-d0203eae1351\" (UID: \"6164de25-5015-4f67-8b21-d0203eae1351\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.325505 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f84311c2-0fb2-4d15-8272-f245fcbc9957-operator-scripts\") pod \"f84311c2-0fb2-4d15-8272-f245fcbc9957\" (UID: \"f84311c2-0fb2-4d15-8272-f245fcbc9957\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.325552 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ade20871-d4bc-4a8f-9635-d28bbae44421-operator-scripts\") pod \"ade20871-d4bc-4a8f-9635-d28bbae44421\" (UID: \"ade20871-d4bc-4a8f-9635-d28bbae44421\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.325617 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjzzf\" (UniqueName: \"kubernetes.io/projected/fffd6b7f-d18d-4ff8-a446-032de1c3e288-kube-api-access-kjzzf\") pod \"fffd6b7f-d18d-4ff8-a446-032de1c3e288\" (UID: \"fffd6b7f-d18d-4ff8-a446-032de1c3e288\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.325648 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fffd6b7f-d18d-4ff8-a446-032de1c3e288-operator-scripts\") pod \"fffd6b7f-d18d-4ff8-a446-032de1c3e288\" (UID: \"fffd6b7f-d18d-4ff8-a446-032de1c3e288\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.325712 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mql5w\" (UniqueName: \"kubernetes.io/projected/6164de25-5015-4f67-8b21-d0203eae1351-kube-api-access-mql5w\") pod \"6164de25-5015-4f67-8b21-d0203eae1351\" (UID: \"6164de25-5015-4f67-8b21-d0203eae1351\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.327521 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f84311c2-0fb2-4d15-8272-f245fcbc9957-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f84311c2-0fb2-4d15-8272-f245fcbc9957" (UID: "f84311c2-0fb2-4d15-8272-f245fcbc9957"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.328074 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d7cf-account-create-update-v5gw9" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.328299 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ade20871-d4bc-4a8f-9635-d28bbae44421-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ade20871-d4bc-4a8f-9635-d28bbae44421" (UID: "ade20871-d4bc-4a8f-9635-d28bbae44421"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.328635 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fffd6b7f-d18d-4ff8-a446-032de1c3e288-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fffd6b7f-d18d-4ff8-a446-032de1c3e288" (UID: "fffd6b7f-d18d-4ff8-a446-032de1c3e288"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.331339 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6164de25-5015-4f67-8b21-d0203eae1351-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6164de25-5015-4f67-8b21-d0203eae1351" (UID: "6164de25-5015-4f67-8b21-d0203eae1351"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.334151 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fffd6b7f-d18d-4ff8-a446-032de1c3e288-kube-api-access-kjzzf" (OuterVolumeSpecName: "kube-api-access-kjzzf") pod "fffd6b7f-d18d-4ff8-a446-032de1c3e288" (UID: "fffd6b7f-d18d-4ff8-a446-032de1c3e288"). InnerVolumeSpecName "kube-api-access-kjzzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.334930 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6164de25-5015-4f67-8b21-d0203eae1351-kube-api-access-mql5w" (OuterVolumeSpecName: "kube-api-access-mql5w") pod "6164de25-5015-4f67-8b21-d0203eae1351" (UID: "6164de25-5015-4f67-8b21-d0203eae1351"). InnerVolumeSpecName "kube-api-access-mql5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.335713 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f84311c2-0fb2-4d15-8272-f245fcbc9957-kube-api-access-tlgrm" (OuterVolumeSpecName: "kube-api-access-tlgrm") pod "f84311c2-0fb2-4d15-8272-f245fcbc9957" (UID: "f84311c2-0fb2-4d15-8272-f245fcbc9957"). InnerVolumeSpecName "kube-api-access-tlgrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.366343 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ade20871-d4bc-4a8f-9635-d28bbae44421-kube-api-access-hz8wn" (OuterVolumeSpecName: "kube-api-access-hz8wn") pod "ade20871-d4bc-4a8f-9635-d28bbae44421" (UID: "ade20871-d4bc-4a8f-9635-d28bbae44421"). InnerVolumeSpecName "kube-api-access-hz8wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.431116 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csqn5\" (UniqueName: \"kubernetes.io/projected/8f31cf24-9bb2-41c6-a825-c5481a27ffb0-kube-api-access-csqn5\") pod \"8f31cf24-9bb2-41c6-a825-c5481a27ffb0\" (UID: \"8f31cf24-9bb2-41c6-a825-c5481a27ffb0\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.431185 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-68wgk\" (UniqueName: \"kubernetes.io/projected/caf815ba-1554-4c46-bc4d-38b8106cced5-kube-api-access-68wgk\") pod \"caf815ba-1554-4c46-bc4d-38b8106cced5\" (UID: \"caf815ba-1554-4c46-bc4d-38b8106cced5\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.431217 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95e1aca8-117b-43da-b0ce-3b132c4a6c60-operator-scripts\") pod \"95e1aca8-117b-43da-b0ce-3b132c4a6c60\" (UID: \"95e1aca8-117b-43da-b0ce-3b132c4a6c60\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.431259 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f31cf24-9bb2-41c6-a825-c5481a27ffb0-operator-scripts\") pod \"8f31cf24-9bb2-41c6-a825-c5481a27ffb0\" (UID: \"8f31cf24-9bb2-41c6-a825-c5481a27ffb0\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.431528 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qd9tj\" (UniqueName: \"kubernetes.io/projected/95e1aca8-117b-43da-b0ce-3b132c4a6c60-kube-api-access-qd9tj\") pod \"95e1aca8-117b-43da-b0ce-3b132c4a6c60\" (UID: \"95e1aca8-117b-43da-b0ce-3b132c4a6c60\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.432923 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caf815ba-1554-4c46-bc4d-38b8106cced5-operator-scripts\") pod \"caf815ba-1554-4c46-bc4d-38b8106cced5\" (UID: \"caf815ba-1554-4c46-bc4d-38b8106cced5\") " Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.434178 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjzzf\" (UniqueName: \"kubernetes.io/projected/fffd6b7f-d18d-4ff8-a446-032de1c3e288-kube-api-access-kjzzf\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.434200 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fffd6b7f-d18d-4ff8-a446-032de1c3e288-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.434210 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mql5w\" (UniqueName: \"kubernetes.io/projected/6164de25-5015-4f67-8b21-d0203eae1351-kube-api-access-mql5w\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.434221 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tlgrm\" (UniqueName: \"kubernetes.io/projected/f84311c2-0fb2-4d15-8272-f245fcbc9957-kube-api-access-tlgrm\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.434231 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hz8wn\" (UniqueName: \"kubernetes.io/projected/ade20871-d4bc-4a8f-9635-d28bbae44421-kube-api-access-hz8wn\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.434240 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6164de25-5015-4f67-8b21-d0203eae1351-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.434249 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f84311c2-0fb2-4d15-8272-f245fcbc9957-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.434257 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ade20871-d4bc-4a8f-9635-d28bbae44421-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.434454 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/95e1aca8-117b-43da-b0ce-3b132c4a6c60-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "95e1aca8-117b-43da-b0ce-3b132c4a6c60" (UID: "95e1aca8-117b-43da-b0ce-3b132c4a6c60"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.434699 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caf815ba-1554-4c46-bc4d-38b8106cced5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "caf815ba-1554-4c46-bc4d-38b8106cced5" (UID: "caf815ba-1554-4c46-bc4d-38b8106cced5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.434857 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f31cf24-9bb2-41c6-a825-c5481a27ffb0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8f31cf24-9bb2-41c6-a825-c5481a27ffb0" (UID: "8f31cf24-9bb2-41c6-a825-c5481a27ffb0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.439200 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95e1aca8-117b-43da-b0ce-3b132c4a6c60-kube-api-access-qd9tj" (OuterVolumeSpecName: "kube-api-access-qd9tj") pod "95e1aca8-117b-43da-b0ce-3b132c4a6c60" (UID: "95e1aca8-117b-43da-b0ce-3b132c4a6c60"). InnerVolumeSpecName "kube-api-access-qd9tj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.441458 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f31cf24-9bb2-41c6-a825-c5481a27ffb0-kube-api-access-csqn5" (OuterVolumeSpecName: "kube-api-access-csqn5") pod "8f31cf24-9bb2-41c6-a825-c5481a27ffb0" (UID: "8f31cf24-9bb2-41c6-a825-c5481a27ffb0"). InnerVolumeSpecName "kube-api-access-csqn5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.446969 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caf815ba-1554-4c46-bc4d-38b8106cced5-kube-api-access-68wgk" (OuterVolumeSpecName: "kube-api-access-68wgk") pod "caf815ba-1554-4c46-bc4d-38b8106cced5" (UID: "caf815ba-1554-4c46-bc4d-38b8106cced5"). InnerVolumeSpecName "kube-api-access-68wgk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.536518 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qd9tj\" (UniqueName: \"kubernetes.io/projected/95e1aca8-117b-43da-b0ce-3b132c4a6c60-kube-api-access-qd9tj\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.536550 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/caf815ba-1554-4c46-bc4d-38b8106cced5-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.536572 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csqn5\" (UniqueName: \"kubernetes.io/projected/8f31cf24-9bb2-41c6-a825-c5481a27ffb0-kube-api-access-csqn5\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.536585 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-68wgk\" (UniqueName: \"kubernetes.io/projected/caf815ba-1554-4c46-bc4d-38b8106cced5-kube-api-access-68wgk\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.536596 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95e1aca8-117b-43da-b0ce-3b132c4a6c60-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.536607 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8f31cf24-9bb2-41c6-a825-c5481a27ffb0-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.856854 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vr4gs" event={"ID":"d0127857-fa77-4179-ad95-7dfb2585169a","Type":"ContainerStarted","Data":"c4033b1c387e1729a85fe822e90af40e36d600af1237c3dd5bfa0f7a2e713838"} Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.873262 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-0b34-account-create-update-grx5t" event={"ID":"caf815ba-1554-4c46-bc4d-38b8106cced5","Type":"ContainerDied","Data":"f8ea5ffcdc876b8a09c5e0549151c76fae75855ddb818a1dd2ad1ffbeb406e04"} Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.873312 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8ea5ffcdc876b8a09c5e0549151c76fae75855ddb818a1dd2ad1ffbeb406e04" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.873371 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-0b34-account-create-update-grx5t" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.890156 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-lgr8r" event={"ID":"fffd6b7f-d18d-4ff8-a446-032de1c3e288","Type":"ContainerDied","Data":"1d6f40093a1a3c19311515e1fff793c70eb3a2dfa93434ab89887f6bd4925f55"} Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.890192 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d6f40093a1a3c19311515e1fff793c70eb3a2dfa93434ab89887f6bd4925f55" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.890251 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-lgr8r" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.892107 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zqbhp" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.892544 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zqbhp" event={"ID":"f84311c2-0fb2-4d15-8272-f245fcbc9957","Type":"ContainerDied","Data":"f0ba084bbb6777a6f08f4f6a9b73305a118dd6797a2e7d80c6e1399e46f2e08e"} Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.892599 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f0ba084bbb6777a6f08f4f6a9b73305a118dd6797a2e7d80c6e1399e46f2e08e" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.893659 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-pbbn5" event={"ID":"95e1aca8-117b-43da-b0ce-3b132c4a6c60","Type":"ContainerDied","Data":"5541690f3f90095c1ef96b23733bf3c7c0eae169eacc7e07f0762b2abd83f741"} Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.893686 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5541690f3f90095c1ef96b23733bf3c7c0eae169eacc7e07f0762b2abd83f741" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.893734 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-pbbn5" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.898984 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-728c-account-create-update-q5spq" event={"ID":"6164de25-5015-4f67-8b21-d0203eae1351","Type":"ContainerDied","Data":"420033fd92e9fa763c3761ecdd60c58d9a8f0110bdf9ffa53eb04dc6561cbff0"} Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.899039 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="420033fd92e9fa763c3761ecdd60c58d9a8f0110bdf9ffa53eb04dc6561cbff0" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.899086 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-728c-account-create-update-q5spq" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.902241 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-d7cf-account-create-update-v5gw9" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.902239 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-d7cf-account-create-update-v5gw9" event={"ID":"8f31cf24-9bb2-41c6-a825-c5481a27ffb0","Type":"ContainerDied","Data":"d85325bcd8670a5a5e670698aec9beb3c0f4e8f5b6fe4afe651f2d18e8d4bba0"} Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.902368 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d85325bcd8670a5a5e670698aec9beb3c0f4e8f5b6fe4afe651f2d18e8d4bba0" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.904870 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-bdfd-account-create-update-9lgz8" Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.905118 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-bdfd-account-create-update-9lgz8" event={"ID":"ade20871-d4bc-4a8f-9635-d28bbae44421","Type":"ContainerDied","Data":"5705f3fea3e0c7373e10a0ea8cd33e6463ef59ce1e5b41d4355ffb5cc5635fd7"} Mar 13 09:34:31 crc kubenswrapper[4930]: I0313 09:34:31.905720 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5705f3fea3e0c7373e10a0ea8cd33e6463ef59ce1e5b41d4355ffb5cc5635fd7" Mar 13 09:34:32 crc kubenswrapper[4930]: I0313 09:34:32.282605 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-vr4gs" podStartSLOduration=2.803117966 podStartE2EDuration="8.282583469s" podCreationTimestamp="2026-03-13 09:34:24 +0000 UTC" firstStartedPulling="2026-03-13 09:34:25.609309176 +0000 UTC m=+1306.359223853" lastFinishedPulling="2026-03-13 09:34:31.088774679 +0000 UTC m=+1311.838689356" observedRunningTime="2026-03-13 09:34:31.8898952 +0000 UTC m=+1312.639809897" watchObservedRunningTime="2026-03-13 09:34:32.282583469 +0000 UTC m=+1313.032498166" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.249596 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.327419 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-g4vq8"] Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.328008 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" podUID="f4d00778-de97-4b50-a752-42109b188b11" containerName="dnsmasq-dns" containerID="cri-o://9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7" gracePeriod=10 Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.831566 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.884015 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-config\") pod \"f4d00778-de97-4b50-a752-42109b188b11\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.884142 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frb69\" (UniqueName: \"kubernetes.io/projected/f4d00778-de97-4b50-a752-42109b188b11-kube-api-access-frb69\") pod \"f4d00778-de97-4b50-a752-42109b188b11\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.884210 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-ovsdbserver-sb\") pod \"f4d00778-de97-4b50-a752-42109b188b11\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.884276 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-ovsdbserver-nb\") pod \"f4d00778-de97-4b50-a752-42109b188b11\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.884318 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-dns-svc\") pod \"f4d00778-de97-4b50-a752-42109b188b11\" (UID: \"f4d00778-de97-4b50-a752-42109b188b11\") " Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.893612 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4d00778-de97-4b50-a752-42109b188b11-kube-api-access-frb69" (OuterVolumeSpecName: "kube-api-access-frb69") pod "f4d00778-de97-4b50-a752-42109b188b11" (UID: "f4d00778-de97-4b50-a752-42109b188b11"). InnerVolumeSpecName "kube-api-access-frb69". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.927523 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4d00778-de97-4b50-a752-42109b188b11" containerID="9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7" exitCode=0 Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.927564 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" event={"ID":"f4d00778-de97-4b50-a752-42109b188b11","Type":"ContainerDied","Data":"9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7"} Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.927591 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" event={"ID":"f4d00778-de97-4b50-a752-42109b188b11","Type":"ContainerDied","Data":"73b864a8044621e56716f7bf6195d4e3ebe32abbad9ded54e4bed89fc9fab815"} Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.927606 4930 scope.go:117] "RemoveContainer" containerID="9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.927713 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.939216 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f4d00778-de97-4b50-a752-42109b188b11" (UID: "f4d00778-de97-4b50-a752-42109b188b11"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.959942 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-config" (OuterVolumeSpecName: "config") pod "f4d00778-de97-4b50-a752-42109b188b11" (UID: "f4d00778-de97-4b50-a752-42109b188b11"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.966667 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f4d00778-de97-4b50-a752-42109b188b11" (UID: "f4d00778-de97-4b50-a752-42109b188b11"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.967125 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f4d00778-de97-4b50-a752-42109b188b11" (UID: "f4d00778-de97-4b50-a752-42109b188b11"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.987016 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frb69\" (UniqueName: \"kubernetes.io/projected/f4d00778-de97-4b50-a752-42109b188b11-kube-api-access-frb69\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.987070 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.987082 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.987095 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:33 crc kubenswrapper[4930]: I0313 09:34:33.987106 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f4d00778-de97-4b50-a752-42109b188b11-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:34 crc kubenswrapper[4930]: I0313 09:34:34.004060 4930 scope.go:117] "RemoveContainer" containerID="52428e5ba5283d2d616708463dc9171cdbca2df3692703b0a5d6fc5413f59f49" Mar 13 09:34:34 crc kubenswrapper[4930]: I0313 09:34:34.030222 4930 scope.go:117] "RemoveContainer" containerID="9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7" Mar 13 09:34:34 crc kubenswrapper[4930]: E0313 09:34:34.030816 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7\": container with ID starting with 9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7 not found: ID does not exist" containerID="9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7" Mar 13 09:34:34 crc kubenswrapper[4930]: I0313 09:34:34.030872 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7"} err="failed to get container status \"9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7\": rpc error: code = NotFound desc = could not find container \"9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7\": container with ID starting with 9f838beb23c7b5636ebfa371436574542b85907307da233c4d3efbf69f7295a7 not found: ID does not exist" Mar 13 09:34:34 crc kubenswrapper[4930]: I0313 09:34:34.030903 4930 scope.go:117] "RemoveContainer" containerID="52428e5ba5283d2d616708463dc9171cdbca2df3692703b0a5d6fc5413f59f49" Mar 13 09:34:34 crc kubenswrapper[4930]: E0313 09:34:34.031448 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52428e5ba5283d2d616708463dc9171cdbca2df3692703b0a5d6fc5413f59f49\": container with ID starting with 52428e5ba5283d2d616708463dc9171cdbca2df3692703b0a5d6fc5413f59f49 not found: ID does not exist" containerID="52428e5ba5283d2d616708463dc9171cdbca2df3692703b0a5d6fc5413f59f49" Mar 13 09:34:34 crc kubenswrapper[4930]: I0313 09:34:34.031487 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52428e5ba5283d2d616708463dc9171cdbca2df3692703b0a5d6fc5413f59f49"} err="failed to get container status \"52428e5ba5283d2d616708463dc9171cdbca2df3692703b0a5d6fc5413f59f49\": rpc error: code = NotFound desc = could not find container \"52428e5ba5283d2d616708463dc9171cdbca2df3692703b0a5d6fc5413f59f49\": container with ID starting with 52428e5ba5283d2d616708463dc9171cdbca2df3692703b0a5d6fc5413f59f49 not found: ID does not exist" Mar 13 09:34:34 crc kubenswrapper[4930]: I0313 09:34:34.252803 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-g4vq8"] Mar 13 09:34:34 crc kubenswrapper[4930]: I0313 09:34:34.261922 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-g4vq8"] Mar 13 09:34:35 crc kubenswrapper[4930]: I0313 09:34:35.950045 4930 generic.go:334] "Generic (PLEG): container finished" podID="d0127857-fa77-4179-ad95-7dfb2585169a" containerID="c4033b1c387e1729a85fe822e90af40e36d600af1237c3dd5bfa0f7a2e713838" exitCode=0 Mar 13 09:34:35 crc kubenswrapper[4930]: I0313 09:34:35.950155 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vr4gs" event={"ID":"d0127857-fa77-4179-ad95-7dfb2585169a","Type":"ContainerDied","Data":"c4033b1c387e1729a85fe822e90af40e36d600af1237c3dd5bfa0f7a2e713838"} Mar 13 09:34:35 crc kubenswrapper[4930]: I0313 09:34:35.986211 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4d00778-de97-4b50-a752-42109b188b11" path="/var/lib/kubelet/pods/f4d00778-de97-4b50-a752-42109b188b11/volumes" Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.336768 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.350521 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0127857-fa77-4179-ad95-7dfb2585169a-config-data\") pod \"d0127857-fa77-4179-ad95-7dfb2585169a\" (UID: \"d0127857-fa77-4179-ad95-7dfb2585169a\") " Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.350648 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkpzh\" (UniqueName: \"kubernetes.io/projected/d0127857-fa77-4179-ad95-7dfb2585169a-kube-api-access-dkpzh\") pod \"d0127857-fa77-4179-ad95-7dfb2585169a\" (UID: \"d0127857-fa77-4179-ad95-7dfb2585169a\") " Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.350790 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0127857-fa77-4179-ad95-7dfb2585169a-combined-ca-bundle\") pod \"d0127857-fa77-4179-ad95-7dfb2585169a\" (UID: \"d0127857-fa77-4179-ad95-7dfb2585169a\") " Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.376674 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0127857-fa77-4179-ad95-7dfb2585169a-kube-api-access-dkpzh" (OuterVolumeSpecName: "kube-api-access-dkpzh") pod "d0127857-fa77-4179-ad95-7dfb2585169a" (UID: "d0127857-fa77-4179-ad95-7dfb2585169a"). InnerVolumeSpecName "kube-api-access-dkpzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.398729 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0127857-fa77-4179-ad95-7dfb2585169a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d0127857-fa77-4179-ad95-7dfb2585169a" (UID: "d0127857-fa77-4179-ad95-7dfb2585169a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.426086 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d0127857-fa77-4179-ad95-7dfb2585169a-config-data" (OuterVolumeSpecName: "config-data") pod "d0127857-fa77-4179-ad95-7dfb2585169a" (UID: "d0127857-fa77-4179-ad95-7dfb2585169a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.453730 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d0127857-fa77-4179-ad95-7dfb2585169a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.453960 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d0127857-fa77-4179-ad95-7dfb2585169a-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.454077 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkpzh\" (UniqueName: \"kubernetes.io/projected/d0127857-fa77-4179-ad95-7dfb2585169a-kube-api-access-dkpzh\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.971854 4930 generic.go:334] "Generic (PLEG): container finished" podID="bf10d12f-578b-400e-b480-60a7343bc344" containerID="9cd8e9f98e28ab2aa6c6fa08257cd36d1c89eaf1805033449f57940002c31365" exitCode=0 Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.973855 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-vr4gs" Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.993304 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g82c9" event={"ID":"bf10d12f-578b-400e-b480-60a7343bc344","Type":"ContainerDied","Data":"9cd8e9f98e28ab2aa6c6fa08257cd36d1c89eaf1805033449f57940002c31365"} Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.993608 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-vr4gs" event={"ID":"d0127857-fa77-4179-ad95-7dfb2585169a","Type":"ContainerDied","Data":"0a4bfca440c9a81c9dbe77e508b4139614be177743a1d5bd23f0a45ed621cca8"} Mar 13 09:34:37 crc kubenswrapper[4930]: I0313 09:34:37.993696 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a4bfca440c9a81c9dbe77e508b4139614be177743a1d5bd23f0a45ed621cca8" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.240724 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b868669f-v422s"] Mar 13 09:34:38 crc kubenswrapper[4930]: E0313 09:34:38.241580 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d00778-de97-4b50-a752-42109b188b11" containerName="dnsmasq-dns" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.241602 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d00778-de97-4b50-a752-42109b188b11" containerName="dnsmasq-dns" Mar 13 09:34:38 crc kubenswrapper[4930]: E0313 09:34:38.241616 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4d00778-de97-4b50-a752-42109b188b11" containerName="init" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.241624 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4d00778-de97-4b50-a752-42109b188b11" containerName="init" Mar 13 09:34:38 crc kubenswrapper[4930]: E0313 09:34:38.241639 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95e1aca8-117b-43da-b0ce-3b132c4a6c60" containerName="mariadb-database-create" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.241647 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="95e1aca8-117b-43da-b0ce-3b132c4a6c60" containerName="mariadb-database-create" Mar 13 09:34:38 crc kubenswrapper[4930]: E0313 09:34:38.241674 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ade20871-d4bc-4a8f-9635-d28bbae44421" containerName="mariadb-account-create-update" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.241682 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ade20871-d4bc-4a8f-9635-d28bbae44421" containerName="mariadb-account-create-update" Mar 13 09:34:38 crc kubenswrapper[4930]: E0313 09:34:38.241697 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0127857-fa77-4179-ad95-7dfb2585169a" containerName="keystone-db-sync" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.241707 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0127857-fa77-4179-ad95-7dfb2585169a" containerName="keystone-db-sync" Mar 13 09:34:38 crc kubenswrapper[4930]: E0313 09:34:38.241724 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f31cf24-9bb2-41c6-a825-c5481a27ffb0" containerName="mariadb-account-create-update" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.241732 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f31cf24-9bb2-41c6-a825-c5481a27ffb0" containerName="mariadb-account-create-update" Mar 13 09:34:38 crc kubenswrapper[4930]: E0313 09:34:38.241750 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f84311c2-0fb2-4d15-8272-f245fcbc9957" containerName="mariadb-database-create" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.241758 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f84311c2-0fb2-4d15-8272-f245fcbc9957" containerName="mariadb-database-create" Mar 13 09:34:38 crc kubenswrapper[4930]: E0313 09:34:38.241773 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86a246ca-6999-4a55-aa0b-b723006d6305" containerName="mariadb-database-create" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.241782 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="86a246ca-6999-4a55-aa0b-b723006d6305" containerName="mariadb-database-create" Mar 13 09:34:38 crc kubenswrapper[4930]: E0313 09:34:38.241796 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fffd6b7f-d18d-4ff8-a446-032de1c3e288" containerName="mariadb-database-create" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.241803 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="fffd6b7f-d18d-4ff8-a446-032de1c3e288" containerName="mariadb-database-create" Mar 13 09:34:38 crc kubenswrapper[4930]: E0313 09:34:38.241815 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf815ba-1554-4c46-bc4d-38b8106cced5" containerName="mariadb-account-create-update" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.241823 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf815ba-1554-4c46-bc4d-38b8106cced5" containerName="mariadb-account-create-update" Mar 13 09:34:38 crc kubenswrapper[4930]: E0313 09:34:38.241835 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6164de25-5015-4f67-8b21-d0203eae1351" containerName="mariadb-account-create-update" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.241842 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6164de25-5015-4f67-8b21-d0203eae1351" containerName="mariadb-account-create-update" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.242063 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6164de25-5015-4f67-8b21-d0203eae1351" containerName="mariadb-account-create-update" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.242079 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="95e1aca8-117b-43da-b0ce-3b132c4a6c60" containerName="mariadb-database-create" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.242090 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ade20871-d4bc-4a8f-9635-d28bbae44421" containerName="mariadb-account-create-update" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.242105 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="caf815ba-1554-4c46-bc4d-38b8106cced5" containerName="mariadb-account-create-update" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.242119 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="fffd6b7f-d18d-4ff8-a446-032de1c3e288" containerName="mariadb-database-create" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.242127 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f84311c2-0fb2-4d15-8272-f245fcbc9957" containerName="mariadb-database-create" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.242139 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f31cf24-9bb2-41c6-a825-c5481a27ffb0" containerName="mariadb-account-create-update" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.242151 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0127857-fa77-4179-ad95-7dfb2585169a" containerName="keystone-db-sync" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.242165 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="86a246ca-6999-4a55-aa0b-b723006d6305" containerName="mariadb-database-create" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.242178 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4d00778-de97-4b50-a752-42109b188b11" containerName="dnsmasq-dns" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.243509 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.269053 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-dns-svc\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.269107 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.269247 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-config\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.269467 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.269646 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nb8t2\" (UniqueName: \"kubernetes.io/projected/2cea6458-de73-4188-b4e5-6f8453fc289b-kube-api-access-nb8t2\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.269696 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.272095 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-v422s"] Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.287846 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-ftp8x"] Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.289191 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.293706 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.299819 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.299977 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ws54h" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.299987 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.300043 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.320241 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ftp8x"] Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.380290 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-config\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.380405 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.380477 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nb8t2\" (UniqueName: \"kubernetes.io/projected/2cea6458-de73-4188-b4e5-6f8453fc289b-kube-api-access-nb8t2\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.380500 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.380578 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-dns-svc\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.380611 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.381639 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-ovsdbserver-nb\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.382292 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-ovsdbserver-sb\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.382646 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-dns-swift-storage-0\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.383615 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-dns-svc\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.390194 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-config\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.459798 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nb8t2\" (UniqueName: \"kubernetes.io/projected/2cea6458-de73-4188-b4e5-6f8453fc289b-kube-api-access-nb8t2\") pod \"dnsmasq-dns-5b868669f-v422s\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.485601 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4txrl\" (UniqueName: \"kubernetes.io/projected/ba1aad44-386d-4a67-8949-7a465e65ea6e-kube-api-access-4txrl\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.485692 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-fernet-keys\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.485746 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-combined-ca-bundle\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.485765 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-credential-keys\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.485830 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-scripts\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.485862 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-config-data\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.534269 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-p4nk9"] Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.535747 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-p4nk9" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.545929 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.546156 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-tbzqx" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.561198 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.568398 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-p4nk9"] Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.588017 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-fernet-keys\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.588107 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-combined-ca-bundle\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.588148 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-credential-keys\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.588238 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-scripts\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.588301 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-config-data\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.588382 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4txrl\" (UniqueName: \"kubernetes.io/projected/ba1aad44-386d-4a67-8949-7a465e65ea6e-kube-api-access-4txrl\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.595701 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-credential-keys\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.608192 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-fernet-keys\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.609993 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-config-data\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.614782 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-scripts\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.618146 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-combined-ca-bundle\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.623362 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4txrl\" (UniqueName: \"kubernetes.io/projected/ba1aad44-386d-4a67-8949-7a465e65ea6e-kube-api-access-4txrl\") pod \"keystone-bootstrap-ftp8x\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.664586 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-g4vq8" podUID="f4d00778-de97-4b50-a752-42109b188b11" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.154:5353: i/o timeout" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.682098 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-h9bf6"] Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.683385 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.697211 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9m4t\" (UniqueName: \"kubernetes.io/projected/ae61cc80-b8f2-401c-87db-5728d566b288-kube-api-access-d9m4t\") pod \"heat-db-sync-p4nk9\" (UID: \"ae61cc80-b8f2-401c-87db-5728d566b288\") " pod="openstack/heat-db-sync-p4nk9" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.697349 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae61cc80-b8f2-401c-87db-5728d566b288-config-data\") pod \"heat-db-sync-p4nk9\" (UID: \"ae61cc80-b8f2-401c-87db-5728d566b288\") " pod="openstack/heat-db-sync-p4nk9" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.697380 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae61cc80-b8f2-401c-87db-5728d566b288-combined-ca-bundle\") pod \"heat-db-sync-p4nk9\" (UID: \"ae61cc80-b8f2-401c-87db-5728d566b288\") " pod="openstack/heat-db-sync-p4nk9" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.697994 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.710778 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-c5wbx" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.740569 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-h9bf6"] Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.802950 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fb676a-c7df-4e03-a4ad-41da78bde051-combined-ca-bundle\") pod \"barbican-db-sync-h9bf6\" (UID: \"d3fb676a-c7df-4e03-a4ad-41da78bde051\") " pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.803325 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9m4t\" (UniqueName: \"kubernetes.io/projected/ae61cc80-b8f2-401c-87db-5728d566b288-kube-api-access-d9m4t\") pod \"heat-db-sync-p4nk9\" (UID: \"ae61cc80-b8f2-401c-87db-5728d566b288\") " pod="openstack/heat-db-sync-p4nk9" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.803414 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae61cc80-b8f2-401c-87db-5728d566b288-config-data\") pod \"heat-db-sync-p4nk9\" (UID: \"ae61cc80-b8f2-401c-87db-5728d566b288\") " pod="openstack/heat-db-sync-p4nk9" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.803459 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae61cc80-b8f2-401c-87db-5728d566b288-combined-ca-bundle\") pod \"heat-db-sync-p4nk9\" (UID: \"ae61cc80-b8f2-401c-87db-5728d566b288\") " pod="openstack/heat-db-sync-p4nk9" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.803514 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pmwb\" (UniqueName: \"kubernetes.io/projected/d3fb676a-c7df-4e03-a4ad-41da78bde051-kube-api-access-4pmwb\") pod \"barbican-db-sync-h9bf6\" (UID: \"d3fb676a-c7df-4e03-a4ad-41da78bde051\") " pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.803547 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3fb676a-c7df-4e03-a4ad-41da78bde051-db-sync-config-data\") pod \"barbican-db-sync-h9bf6\" (UID: \"d3fb676a-c7df-4e03-a4ad-41da78bde051\") " pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.832853 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-pdp9j"] Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.834259 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.834363 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae61cc80-b8f2-401c-87db-5728d566b288-config-data\") pod \"heat-db-sync-p4nk9\" (UID: \"ae61cc80-b8f2-401c-87db-5728d566b288\") " pod="openstack/heat-db-sync-p4nk9" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.843486 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-l72lb" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.843792 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.843967 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.862118 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae61cc80-b8f2-401c-87db-5728d566b288-combined-ca-bundle\") pod \"heat-db-sync-p4nk9\" (UID: \"ae61cc80-b8f2-401c-87db-5728d566b288\") " pod="openstack/heat-db-sync-p4nk9" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.874324 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9m4t\" (UniqueName: \"kubernetes.io/projected/ae61cc80-b8f2-401c-87db-5728d566b288-kube-api-access-d9m4t\") pod \"heat-db-sync-p4nk9\" (UID: \"ae61cc80-b8f2-401c-87db-5728d566b288\") " pod="openstack/heat-db-sync-p4nk9" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.896402 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-pdp9j"] Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.910494 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pmwb\" (UniqueName: \"kubernetes.io/projected/d3fb676a-c7df-4e03-a4ad-41da78bde051-kube-api-access-4pmwb\") pod \"barbican-db-sync-h9bf6\" (UID: \"d3fb676a-c7df-4e03-a4ad-41da78bde051\") " pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.910549 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3fb676a-c7df-4e03-a4ad-41da78bde051-db-sync-config-data\") pod \"barbican-db-sync-h9bf6\" (UID: \"d3fb676a-c7df-4e03-a4ad-41da78bde051\") " pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.910635 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf2b4\" (UniqueName: \"kubernetes.io/projected/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-kube-api-access-hf2b4\") pod \"neutron-db-sync-pdp9j\" (UID: \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\") " pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.910667 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fb676a-c7df-4e03-a4ad-41da78bde051-combined-ca-bundle\") pod \"barbican-db-sync-h9bf6\" (UID: \"d3fb676a-c7df-4e03-a4ad-41da78bde051\") " pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.910699 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-config\") pod \"neutron-db-sync-pdp9j\" (UID: \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\") " pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.910836 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-combined-ca-bundle\") pod \"neutron-db-sync-pdp9j\" (UID: \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\") " pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.921684 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.978052 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fb676a-c7df-4e03-a4ad-41da78bde051-combined-ca-bundle\") pod \"barbican-db-sync-h9bf6\" (UID: \"d3fb676a-c7df-4e03-a4ad-41da78bde051\") " pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.992945 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pmwb\" (UniqueName: \"kubernetes.io/projected/d3fb676a-c7df-4e03-a4ad-41da78bde051-kube-api-access-4pmwb\") pod \"barbican-db-sync-h9bf6\" (UID: \"d3fb676a-c7df-4e03-a4ad-41da78bde051\") " pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:34:38 crc kubenswrapper[4930]: I0313 09:34:38.994945 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3fb676a-c7df-4e03-a4ad-41da78bde051-db-sync-config-data\") pod \"barbican-db-sync-h9bf6\" (UID: \"d3fb676a-c7df-4e03-a4ad-41da78bde051\") " pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.001037 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-7dsrx"] Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.002327 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.008668 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.012883 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf2b4\" (UniqueName: \"kubernetes.io/projected/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-kube-api-access-hf2b4\") pod \"neutron-db-sync-pdp9j\" (UID: \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\") " pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.012957 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-p6m7w" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.012992 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-config\") pod \"neutron-db-sync-pdp9j\" (UID: \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\") " pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.013018 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-scripts\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.013729 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-combined-ca-bundle\") pod \"neutron-db-sync-pdp9j\" (UID: \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\") " pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.013827 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-config-data\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.013859 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-combined-ca-bundle\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.013879 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmcpt\" (UniqueName: \"kubernetes.io/projected/98c022d1-9a39-4f3e-8108-e4be2b287077-kube-api-access-gmcpt\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.013953 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/98c022d1-9a39-4f3e-8108-e4be2b287077-etc-machine-id\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.013997 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-db-sync-config-data\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.021844 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.044450 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7dsrx"] Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.046328 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-config\") pod \"neutron-db-sync-pdp9j\" (UID: \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\") " pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.051629 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-combined-ca-bundle\") pod \"neutron-db-sync-pdp9j\" (UID: \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\") " pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.070083 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf2b4\" (UniqueName: \"kubernetes.io/projected/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-kube-api-access-hf2b4\") pod \"neutron-db-sync-pdp9j\" (UID: \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\") " pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.073901 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-djnqj"] Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.075653 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.080314 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-z6lh5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.080564 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.081067 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.089860 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.100489 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-v422s"] Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.115941 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-scripts\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.116149 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-config-data\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.116276 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-combined-ca-bundle\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.116377 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmcpt\" (UniqueName: \"kubernetes.io/projected/98c022d1-9a39-4f3e-8108-e4be2b287077-kube-api-access-gmcpt\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.116519 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/98c022d1-9a39-4f3e-8108-e4be2b287077-etc-machine-id\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.130068 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-db-sync-config-data\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.121084 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-djnqj"] Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.129174 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-combined-ca-bundle\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.121332 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/98c022d1-9a39-4f3e-8108-e4be2b287077-etc-machine-id\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.135991 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-scripts\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.136814 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-db-sync-config-data\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.142388 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-config-data\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.152384 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmcpt\" (UniqueName: \"kubernetes.io/projected/98c022d1-9a39-4f3e-8108-e4be2b287077-kube-api-access-gmcpt\") pod \"cinder-db-sync-7dsrx\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.160825 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-hsrk5"] Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.162515 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.163743 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-p4nk9" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.182595 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-hsrk5"] Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.214270 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.221773 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.234224 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-config\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.234267 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.234295 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cktrp\" (UniqueName: \"kubernetes.io/projected/1da55f1c-1e80-4231-9da7-3a9ca0923b32-kube-api-access-cktrp\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.234326 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-combined-ca-bundle\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.234860 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-logs\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.234907 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.234980 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqwxf\" (UniqueName: \"kubernetes.io/projected/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-kube-api-access-bqwxf\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.235004 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-dns-svc\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.235122 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-scripts\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.235172 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.235365 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-config-data\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.337031 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqwxf\" (UniqueName: \"kubernetes.io/projected/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-kube-api-access-bqwxf\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.337081 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-dns-svc\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.337142 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-scripts\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.337177 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.337214 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-config-data\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.337749 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-config\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.337818 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.337875 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cktrp\" (UniqueName: \"kubernetes.io/projected/1da55f1c-1e80-4231-9da7-3a9ca0923b32-kube-api-access-cktrp\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.337959 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-combined-ca-bundle\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.338059 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-logs\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.338143 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.339198 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-ovsdbserver-sb\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.340068 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-dns-swift-storage-0\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.340752 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-config\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.341849 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-logs\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.340068 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-ovsdbserver-nb\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.343508 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-dns-svc\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.348056 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-combined-ca-bundle\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.348793 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-scripts\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.356978 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-config-data\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.371183 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cktrp\" (UniqueName: \"kubernetes.io/projected/1da55f1c-1e80-4231-9da7-3a9ca0923b32-kube-api-access-cktrp\") pod \"dnsmasq-dns-cf78879c9-hsrk5\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.371397 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqwxf\" (UniqueName: \"kubernetes.io/projected/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-kube-api-access-bqwxf\") pod \"placement-db-sync-djnqj\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.554038 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-djnqj" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.572786 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.619980 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-v422s"] Mar 13 09:34:39 crc kubenswrapper[4930]: W0313 09:34:39.656199 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2cea6458_de73_4188_b4e5_6f8453fc289b.slice/crio-9fc69ea8e3ae4e9fc0d89ebc74f88db6a5503bf3bd029222af9c7c5906e38c22 WatchSource:0}: Error finding container 9fc69ea8e3ae4e9fc0d89ebc74f88db6a5503bf3bd029222af9c7c5906e38c22: Status 404 returned error can't find the container with id 9fc69ea8e3ae4e9fc0d89ebc74f88db6a5503bf3bd029222af9c7c5906e38c22 Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.739217 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.773730 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.773853 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.782241 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.782448 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.889818 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-ftp8x"] Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.951198 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc498063-054b-4708-981d-65da056e5b2f-log-httpd\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.951596 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-scripts\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.952458 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc498063-054b-4708-981d-65da056e5b2f-run-httpd\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.952508 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.952625 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wrssc\" (UniqueName: \"kubernetes.io/projected/bc498063-054b-4708-981d-65da056e5b2f-kube-api-access-wrssc\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.952673 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:39 crc kubenswrapper[4930]: I0313 09:34:39.952747 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-config-data\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.023779 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g82c9" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.057521 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.057610 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-config-data\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.057695 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc498063-054b-4708-981d-65da056e5b2f-log-httpd\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.057751 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-scripts\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.057793 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc498063-054b-4708-981d-65da056e5b2f-run-httpd\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.057820 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.057864 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wrssc\" (UniqueName: \"kubernetes.io/projected/bc498063-054b-4708-981d-65da056e5b2f-kube-api-access-wrssc\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.058494 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc498063-054b-4708-981d-65da056e5b2f-log-httpd\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.058653 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc498063-054b-4708-981d-65da056e5b2f-run-httpd\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.061109 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.061207 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.071046 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.072270 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-h9bf6"] Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.072312 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ftp8x" event={"ID":"ba1aad44-386d-4a67-8949-7a465e65ea6e","Type":"ContainerStarted","Data":"d26896a09104b76c54ac90597c5b5fdc99bf9714f2064aef68417ba4ba430773"} Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.075007 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-v422s" event={"ID":"2cea6458-de73-4188-b4e5-6f8453fc289b","Type":"ContainerStarted","Data":"9fc69ea8e3ae4e9fc0d89ebc74f88db6a5503bf3bd029222af9c7c5906e38c22"} Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.083051 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-g82c9" event={"ID":"bf10d12f-578b-400e-b480-60a7343bc344","Type":"ContainerDied","Data":"d2a628d60e282a82ec857647464ea0b39ca58c7e8dfc03ce8c3c8ba0f8212941"} Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.083090 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d2a628d60e282a82ec857647464ea0b39ca58c7e8dfc03ce8c3c8ba0f8212941" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.083139 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-g82c9" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.086758 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-scripts\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.087456 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-config-data\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.098537 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.099084 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wrssc\" (UniqueName: \"kubernetes.io/projected/bc498063-054b-4708-981d-65da056e5b2f-kube-api-access-wrssc\") pod \"ceilometer-0\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.148794 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.160013 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-db-sync-config-data\") pod \"bf10d12f-578b-400e-b480-60a7343bc344\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.160087 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-config-data\") pod \"bf10d12f-578b-400e-b480-60a7343bc344\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.160135 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-combined-ca-bundle\") pod \"bf10d12f-578b-400e-b480-60a7343bc344\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.160225 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95f64\" (UniqueName: \"kubernetes.io/projected/bf10d12f-578b-400e-b480-60a7343bc344-kube-api-access-95f64\") pod \"bf10d12f-578b-400e-b480-60a7343bc344\" (UID: \"bf10d12f-578b-400e-b480-60a7343bc344\") " Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.169363 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bf10d12f-578b-400e-b480-60a7343bc344" (UID: "bf10d12f-578b-400e-b480-60a7343bc344"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.181167 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf10d12f-578b-400e-b480-60a7343bc344-kube-api-access-95f64" (OuterVolumeSpecName: "kube-api-access-95f64") pod "bf10d12f-578b-400e-b480-60a7343bc344" (UID: "bf10d12f-578b-400e-b480-60a7343bc344"). InnerVolumeSpecName "kube-api-access-95f64". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.234340 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-config-data" (OuterVolumeSpecName: "config-data") pod "bf10d12f-578b-400e-b480-60a7343bc344" (UID: "bf10d12f-578b-400e-b480-60a7343bc344"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.243371 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bf10d12f-578b-400e-b480-60a7343bc344" (UID: "bf10d12f-578b-400e-b480-60a7343bc344"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.263121 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95f64\" (UniqueName: \"kubernetes.io/projected/bf10d12f-578b-400e-b480-60a7343bc344-kube-api-access-95f64\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.263236 4930 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.263376 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.263459 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf10d12f-578b-400e-b480-60a7343bc344-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.562718 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-p4nk9"] Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.595240 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-djnqj"] Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.608994 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-pdp9j"] Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.676017 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-hsrk5"] Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.689088 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-7dsrx"] Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.899103 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:34:40 crc kubenswrapper[4930]: W0313 09:34:40.935285 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbc498063_054b_4708_981d_65da056e5b2f.slice/crio-e531cf6f90563bcb0e755872c6ae653ff60002fbbdd0eda5902d1ae27c193589 WatchSource:0}: Error finding container e531cf6f90563bcb0e755872c6ae653ff60002fbbdd0eda5902d1ae27c193589: Status 404 returned error can't find the container with id e531cf6f90563bcb0e755872c6ae653ff60002fbbdd0eda5902d1ae27c193589 Mar 13 09:34:40 crc kubenswrapper[4930]: I0313 09:34:40.994424 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.141324 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-p4nk9" event={"ID":"ae61cc80-b8f2-401c-87db-5728d566b288","Type":"ContainerStarted","Data":"4c2f6d942ca4ca1a6cb8e8011be467e830c740a36c8a97c24b93b563e0766c3f"} Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.145899 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h9bf6" event={"ID":"d3fb676a-c7df-4e03-a4ad-41da78bde051","Type":"ContainerStarted","Data":"c617f8fbba5d3e9be1a4c1e4e7a1c07d1562efcb5c7f5158a487e6b45aa8e2be"} Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.151054 4930 generic.go:334] "Generic (PLEG): container finished" podID="2cea6458-de73-4188-b4e5-6f8453fc289b" containerID="1cacc25266ca59619e086735886ca9a7e203d2cb196fe34b2b47c5941b393429" exitCode=0 Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.151399 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-v422s" event={"ID":"2cea6458-de73-4188-b4e5-6f8453fc289b","Type":"ContainerDied","Data":"1cacc25266ca59619e086735886ca9a7e203d2cb196fe34b2b47c5941b393429"} Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.175729 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc498063-054b-4708-981d-65da056e5b2f","Type":"ContainerStarted","Data":"e531cf6f90563bcb0e755872c6ae653ff60002fbbdd0eda5902d1ae27c193589"} Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.187370 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" event={"ID":"1da55f1c-1e80-4231-9da7-3a9ca0923b32","Type":"ContainerStarted","Data":"3485006de7ca5b97e523f94dc354bd9fe5ac7a3d9a48d7cbd5ce273fdc696875"} Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.206537 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-pdp9j" event={"ID":"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc","Type":"ContainerStarted","Data":"4c5af5cec93f7f8bc461701a1335dac3cef8fe75b0f13dba8493cc37bc39c6bd"} Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.227086 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ftp8x" event={"ID":"ba1aad44-386d-4a67-8949-7a465e65ea6e","Type":"ContainerStarted","Data":"b831a2f73338ffdf8908b03cae19f580a7caad9da82479cc713725ccc6661253"} Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.233885 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7dsrx" event={"ID":"98c022d1-9a39-4f3e-8108-e4be2b287077","Type":"ContainerStarted","Data":"9d82304b36b5b9ad7d85fbaf6cd66db99326cb32098ccb87ecd939ee6cee8456"} Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.244339 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-djnqj" event={"ID":"10e08b2d-bcbf-4b39-9baa-e4ca5213567c","Type":"ContainerStarted","Data":"02a0e5980658806ec9b3ca5294d033a490a0be1174d565afeb30a636c3aa2ef3"} Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.333241 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-ftp8x" podStartSLOduration=3.333217593 podStartE2EDuration="3.333217593s" podCreationTimestamp="2026-03-13 09:34:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:34:41.266081565 +0000 UTC m=+1322.015996252" watchObservedRunningTime="2026-03-13 09:34:41.333217593 +0000 UTC m=+1322.083132270" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.610988 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-hsrk5"] Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.675850 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-g6rkm"] Mar 13 09:34:41 crc kubenswrapper[4930]: E0313 09:34:41.694933 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf10d12f-578b-400e-b480-60a7343bc344" containerName="glance-db-sync" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.694970 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf10d12f-578b-400e-b480-60a7343bc344" containerName="glance-db-sync" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.695716 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf10d12f-578b-400e-b480-60a7343bc344" containerName="glance-db-sync" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.708991 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.769006 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-g6rkm"] Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.837347 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.837471 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-config\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.837563 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.837642 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.837674 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.837716 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dxhlm\" (UniqueName: \"kubernetes.io/projected/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-kube-api-access-dxhlm\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.936990 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.941158 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.941215 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dxhlm\" (UniqueName: \"kubernetes.io/projected/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-kube-api-access-dxhlm\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.941291 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.941350 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-config\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.941406 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.941464 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.942362 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.942919 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.943236 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-config\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.943857 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.946910 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:41 crc kubenswrapper[4930]: I0313 09:34:41.970381 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dxhlm\" (UniqueName: \"kubernetes.io/projected/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-kube-api-access-dxhlm\") pod \"dnsmasq-dns-56df8fb6b7-g6rkm\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.043536 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-ovsdbserver-nb\") pod \"2cea6458-de73-4188-b4e5-6f8453fc289b\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.043615 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-dns-svc\") pod \"2cea6458-de73-4188-b4e5-6f8453fc289b\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.043687 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nb8t2\" (UniqueName: \"kubernetes.io/projected/2cea6458-de73-4188-b4e5-6f8453fc289b-kube-api-access-nb8t2\") pod \"2cea6458-de73-4188-b4e5-6f8453fc289b\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.043833 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-config\") pod \"2cea6458-de73-4188-b4e5-6f8453fc289b\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.043893 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-ovsdbserver-sb\") pod \"2cea6458-de73-4188-b4e5-6f8453fc289b\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.043941 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-dns-swift-storage-0\") pod \"2cea6458-de73-4188-b4e5-6f8453fc289b\" (UID: \"2cea6458-de73-4188-b4e5-6f8453fc289b\") " Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.066682 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cea6458-de73-4188-b4e5-6f8453fc289b-kube-api-access-nb8t2" (OuterVolumeSpecName: "kube-api-access-nb8t2") pod "2cea6458-de73-4188-b4e5-6f8453fc289b" (UID: "2cea6458-de73-4188-b4e5-6f8453fc289b"). InnerVolumeSpecName "kube-api-access-nb8t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.081929 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2cea6458-de73-4188-b4e5-6f8453fc289b" (UID: "2cea6458-de73-4188-b4e5-6f8453fc289b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.111582 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2cea6458-de73-4188-b4e5-6f8453fc289b" (UID: "2cea6458-de73-4188-b4e5-6f8453fc289b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.128917 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2cea6458-de73-4188-b4e5-6f8453fc289b" (UID: "2cea6458-de73-4188-b4e5-6f8453fc289b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.134758 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2cea6458-de73-4188-b4e5-6f8453fc289b" (UID: "2cea6458-de73-4188-b4e5-6f8453fc289b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.146956 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nb8t2\" (UniqueName: \"kubernetes.io/projected/2cea6458-de73-4188-b4e5-6f8453fc289b-kube-api-access-nb8t2\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.146990 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.146999 4930 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.147009 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.147019 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.175774 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-config" (OuterVolumeSpecName: "config") pod "2cea6458-de73-4188-b4e5-6f8453fc289b" (UID: "2cea6458-de73-4188-b4e5-6f8453fc289b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.223112 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.248662 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2cea6458-de73-4188-b4e5-6f8453fc289b-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.273133 4930 generic.go:334] "Generic (PLEG): container finished" podID="1da55f1c-1e80-4231-9da7-3a9ca0923b32" containerID="191dd018bbefe03bb78185dd3accd2667b5261334e58677f35f414ec97581a61" exitCode=0 Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.273320 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" event={"ID":"1da55f1c-1e80-4231-9da7-3a9ca0923b32","Type":"ContainerDied","Data":"191dd018bbefe03bb78185dd3accd2667b5261334e58677f35f414ec97581a61"} Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.281917 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-pdp9j" event={"ID":"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc","Type":"ContainerStarted","Data":"681d0cd0f264d27cd661797cc1039f195824ca91fbb3eac1e6d8422a9da38dc1"} Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.285274 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b868669f-v422s" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.298753 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b868669f-v422s" event={"ID":"2cea6458-de73-4188-b4e5-6f8453fc289b","Type":"ContainerDied","Data":"9fc69ea8e3ae4e9fc0d89ebc74f88db6a5503bf3bd029222af9c7c5906e38c22"} Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.298806 4930 scope.go:117] "RemoveContainer" containerID="1cacc25266ca59619e086735886ca9a7e203d2cb196fe34b2b47c5941b393429" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.307804 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.307844 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.326628 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-pdp9j" podStartSLOduration=4.326602416 podStartE2EDuration="4.326602416s" podCreationTimestamp="2026-03-13 09:34:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:34:42.315176917 +0000 UTC m=+1323.065091604" watchObservedRunningTime="2026-03-13 09:34:42.326602416 +0000 UTC m=+1323.076517093" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.415396 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-v422s"] Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.427585 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b868669f-v422s"] Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.508125 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:34:42 crc kubenswrapper[4930]: E0313 09:34:42.508792 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cea6458-de73-4188-b4e5-6f8453fc289b" containerName="init" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.508807 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cea6458-de73-4188-b4e5-6f8453fc289b" containerName="init" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.509043 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cea6458-de73-4188-b4e5-6f8453fc289b" containerName="init" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.510140 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.514547 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.514715 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-gpm2r" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.523202 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.524303 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.668519 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-scripts\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.668693 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b25b222-49ea-451d-81b8-b9fc7b54459b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.668714 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-config-data\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.668739 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.668772 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.668869 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wssn4\" (UniqueName: \"kubernetes.io/projected/9b25b222-49ea-451d-81b8-b9fc7b54459b-kube-api-access-wssn4\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.668915 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b25b222-49ea-451d-81b8-b9fc7b54459b-logs\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.771058 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-scripts\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.771608 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b25b222-49ea-451d-81b8-b9fc7b54459b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.771640 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-config-data\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.771670 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.771740 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.771799 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wssn4\" (UniqueName: \"kubernetes.io/projected/9b25b222-49ea-451d-81b8-b9fc7b54459b-kube-api-access-wssn4\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.771834 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b25b222-49ea-451d-81b8-b9fc7b54459b-logs\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.772748 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b25b222-49ea-451d-81b8-b9fc7b54459b-logs\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.773669 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b25b222-49ea-451d-81b8-b9fc7b54459b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.776176 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.776212 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/99223c607948bed00b9fe5109399df143422767dfa8a32b71d4b371c3748c0ce/globalmount\"" pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.777239 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-scripts\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.780774 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-config-data\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.786904 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.804498 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.807064 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.809692 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.811664 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wssn4\" (UniqueName: \"kubernetes.io/projected/9b25b222-49ea-451d-81b8-b9fc7b54459b-kube-api-access-wssn4\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: E0313 09:34:42.827578 4930 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Mar 13 09:34:42 crc kubenswrapper[4930]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/1da55f1c-1e80-4231-9da7-3a9ca0923b32/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Mar 13 09:34:42 crc kubenswrapper[4930]: > podSandboxID="3485006de7ca5b97e523f94dc354bd9fe5ac7a3d9a48d7cbd5ce273fdc696875" Mar 13 09:34:42 crc kubenswrapper[4930]: E0313 09:34:42.827736 4930 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 13 09:34:42 crc kubenswrapper[4930]: container &Container{Name:dnsmasq-dns,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5d9h556h556h5cdh98h85h68ch68ch58fh578h57dh5b5h554hd5h7ch68bh68fhf5h59h644h57chc7h648h5bbh684h8fh688h689hd5h679h565h546q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-swift-storage-0,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-swift-storage-0,SubPath:dns-swift-storage-0,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-nb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-nb,SubPath:ovsdbserver-nb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ovsdbserver-sb,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/ovsdbserver-sb,SubPath:ovsdbserver-sb,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-cktrp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 5353 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-cf78879c9-hsrk5_openstack(1da55f1c-1e80-4231-9da7-3a9ca0923b32): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/1da55f1c-1e80-4231-9da7-3a9ca0923b32/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Mar 13 09:34:42 crc kubenswrapper[4930]: > logger="UnhandledError" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.829542 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:34:42 crc kubenswrapper[4930]: E0313 09:34:42.831974 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"dnsmasq-dns\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/1da55f1c-1e80-4231-9da7-3a9ca0923b32/volume-subpaths/dns-svc/dnsmasq-dns/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" podUID="1da55f1c-1e80-4231-9da7-3a9ca0923b32" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.876532 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.876591 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtssp\" (UniqueName: \"kubernetes.io/projected/8cbd73f5-ae52-40dd-9247-ce497789fba1-kube-api-access-gtssp\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.876612 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbd73f5-ae52-40dd-9247-ce497789fba1-logs\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.876642 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.876676 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8cbd73f5-ae52-40dd-9247-ce497789fba1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.876696 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.876822 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.891816 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"glance-default-external-api-0\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.961184 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-g6rkm"] Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.966860 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.979991 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8cbd73f5-ae52-40dd-9247-ce497789fba1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.980042 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.980121 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.980217 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.980255 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtssp\" (UniqueName: \"kubernetes.io/projected/8cbd73f5-ae52-40dd-9247-ce497789fba1-kube-api-access-gtssp\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.980271 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbd73f5-ae52-40dd-9247-ce497789fba1-logs\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.980300 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.981381 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8cbd73f5-ae52-40dd-9247-ce497789fba1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.981999 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbd73f5-ae52-40dd-9247-ce497789fba1-logs\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.983999 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.984040 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a4f4242b5f3665838ee27f4f8b1e1823b22b221bc7d207651abbb8d37832373d/globalmount\"" pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.988606 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:42 crc kubenswrapper[4930]: I0313 09:34:42.989028 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:43 crc kubenswrapper[4930]: I0313 09:34:43.003984 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:43 crc kubenswrapper[4930]: I0313 09:34:43.013207 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtssp\" (UniqueName: \"kubernetes.io/projected/8cbd73f5-ae52-40dd-9247-ce497789fba1-kube-api-access-gtssp\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:43 crc kubenswrapper[4930]: I0313 09:34:43.043278 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"glance-default-internal-api-0\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:34:43 crc kubenswrapper[4930]: I0313 09:34:43.257378 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 09:34:43 crc kubenswrapper[4930]: I0313 09:34:43.354738 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" event={"ID":"d6f38ccc-56eb-4a6e-88ab-321f81d721d5","Type":"ContainerStarted","Data":"3be41492351932246cfe5e38c9ccf6d92ed4ca9e400b8c76579d01c4671732bf"} Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.027473 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2cea6458-de73-4188-b4e5-6f8453fc289b" path="/var/lib/kubelet/pods/2cea6458-de73-4188-b4e5-6f8453fc289b/volumes" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.064150 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.113306 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.125565 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-dns-svc\") pod \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.125645 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-ovsdbserver-nb\") pod \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.125766 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-ovsdbserver-sb\") pod \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.125822 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cktrp\" (UniqueName: \"kubernetes.io/projected/1da55f1c-1e80-4231-9da7-3a9ca0923b32-kube-api-access-cktrp\") pod \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.125949 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-config\") pod \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.126069 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-dns-swift-storage-0\") pod \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\" (UID: \"1da55f1c-1e80-4231-9da7-3a9ca0923b32\") " Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.139360 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1da55f1c-1e80-4231-9da7-3a9ca0923b32-kube-api-access-cktrp" (OuterVolumeSpecName: "kube-api-access-cktrp") pod "1da55f1c-1e80-4231-9da7-3a9ca0923b32" (UID: "1da55f1c-1e80-4231-9da7-3a9ca0923b32"). InnerVolumeSpecName "kube-api-access-cktrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.228472 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cktrp\" (UniqueName: \"kubernetes.io/projected/1da55f1c-1e80-4231-9da7-3a9ca0923b32-kube-api-access-cktrp\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.243226 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.245169 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1da55f1c-1e80-4231-9da7-3a9ca0923b32" (UID: "1da55f1c-1e80-4231-9da7-3a9ca0923b32"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.261455 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1da55f1c-1e80-4231-9da7-3a9ca0923b32" (UID: "1da55f1c-1e80-4231-9da7-3a9ca0923b32"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.271841 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-config" (OuterVolumeSpecName: "config") pod "1da55f1c-1e80-4231-9da7-3a9ca0923b32" (UID: "1da55f1c-1e80-4231-9da7-3a9ca0923b32"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.274803 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1da55f1c-1e80-4231-9da7-3a9ca0923b32" (UID: "1da55f1c-1e80-4231-9da7-3a9ca0923b32"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.277328 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1da55f1c-1e80-4231-9da7-3a9ca0923b32" (UID: "1da55f1c-1e80-4231-9da7-3a9ca0923b32"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.333756 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.333795 4930 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.333807 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.333817 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.333827 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1da55f1c-1e80-4231-9da7-3a9ca0923b32-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.389979 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" event={"ID":"1da55f1c-1e80-4231-9da7-3a9ca0923b32","Type":"ContainerDied","Data":"3485006de7ca5b97e523f94dc354bd9fe5ac7a3d9a48d7cbd5ce273fdc696875"} Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.390063 4930 scope.go:117] "RemoveContainer" containerID="191dd018bbefe03bb78185dd3accd2667b5261334e58677f35f414ec97581a61" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.390230 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cf78879c9-hsrk5" Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.403944 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b25b222-49ea-451d-81b8-b9fc7b54459b","Type":"ContainerStarted","Data":"e0a97ab608d97e88728aa48b517ffcc138e8ca7b991f888a0ae746688d02f3ed"} Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.412107 4930 generic.go:334] "Generic (PLEG): container finished" podID="d6f38ccc-56eb-4a6e-88ab-321f81d721d5" containerID="49c88ce438a76d30e7936c57f380667742b09a6215e9d1d7f78054f1e44091c9" exitCode=0 Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.412168 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" event={"ID":"d6f38ccc-56eb-4a6e-88ab-321f81d721d5","Type":"ContainerDied","Data":"49c88ce438a76d30e7936c57f380667742b09a6215e9d1d7f78054f1e44091c9"} Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.420989 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8cbd73f5-ae52-40dd-9247-ce497789fba1","Type":"ContainerStarted","Data":"3e579889c31aac96e4cb18129fd4cdfb7e39a6b2d4ea24140129d2f4393c104a"} Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.550265 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-hsrk5"] Mar 13 09:34:44 crc kubenswrapper[4930]: I0313 09:34:44.565058 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-cf78879c9-hsrk5"] Mar 13 09:34:45 crc kubenswrapper[4930]: I0313 09:34:45.490692 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b25b222-49ea-451d-81b8-b9fc7b54459b","Type":"ContainerStarted","Data":"46bb7c7db7fb23dff4d800e03e2dc775a416dd3708e10130bbdb8e2654e6a0fb"} Mar 13 09:34:45 crc kubenswrapper[4930]: I0313 09:34:45.498964 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" event={"ID":"d6f38ccc-56eb-4a6e-88ab-321f81d721d5","Type":"ContainerStarted","Data":"dd92ca79209c20efba50eff1a6182a6fde22fa49443435bce612d221c18eb1d0"} Mar 13 09:34:45 crc kubenswrapper[4930]: I0313 09:34:45.499036 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:45 crc kubenswrapper[4930]: I0313 09:34:45.502530 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8cbd73f5-ae52-40dd-9247-ce497789fba1","Type":"ContainerStarted","Data":"02b6cf5f63d74d4e0767c2cf29416755be61982783c7e88f3db5e700b5de4b93"} Mar 13 09:34:45 crc kubenswrapper[4930]: I0313 09:34:45.523532 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" podStartSLOduration=4.52351324 podStartE2EDuration="4.52351324s" podCreationTimestamp="2026-03-13 09:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:34:45.523257763 +0000 UTC m=+1326.273172440" watchObservedRunningTime="2026-03-13 09:34:45.52351324 +0000 UTC m=+1326.273427917" Mar 13 09:34:45 crc kubenswrapper[4930]: I0313 09:34:45.992155 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1da55f1c-1e80-4231-9da7-3a9ca0923b32" path="/var/lib/kubelet/pods/1da55f1c-1e80-4231-9da7-3a9ca0923b32/volumes" Mar 13 09:34:46 crc kubenswrapper[4930]: I0313 09:34:46.524761 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b25b222-49ea-451d-81b8-b9fc7b54459b","Type":"ContainerStarted","Data":"694007dc633201b93f8eaaa97d34fc707f3c2459c2ee7752f35bf2761355d889"} Mar 13 09:34:46 crc kubenswrapper[4930]: I0313 09:34:46.552567 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.552546407 podStartE2EDuration="5.552546407s" podCreationTimestamp="2026-03-13 09:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:34:46.549019135 +0000 UTC m=+1327.298933802" watchObservedRunningTime="2026-03-13 09:34:46.552546407 +0000 UTC m=+1327.302461084" Mar 13 09:34:47 crc kubenswrapper[4930]: I0313 09:34:47.548577 4930 generic.go:334] "Generic (PLEG): container finished" podID="ba1aad44-386d-4a67-8949-7a465e65ea6e" containerID="b831a2f73338ffdf8908b03cae19f580a7caad9da82479cc713725ccc6661253" exitCode=0 Mar 13 09:34:47 crc kubenswrapper[4930]: I0313 09:34:47.550742 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ftp8x" event={"ID":"ba1aad44-386d-4a67-8949-7a465e65ea6e","Type":"ContainerDied","Data":"b831a2f73338ffdf8908b03cae19f580a7caad9da82479cc713725ccc6661253"} Mar 13 09:34:47 crc kubenswrapper[4930]: I0313 09:34:47.558413 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8cbd73f5-ae52-40dd-9247-ce497789fba1","Type":"ContainerStarted","Data":"079475b3d19765c42750282c69df7d259547fab985777833c4bb81e3106533ba"} Mar 13 09:34:47 crc kubenswrapper[4930]: I0313 09:34:47.609577 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=6.609555215 podStartE2EDuration="6.609555215s" podCreationTimestamp="2026-03-13 09:34:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:34:47.600361554 +0000 UTC m=+1328.350276241" watchObservedRunningTime="2026-03-13 09:34:47.609555215 +0000 UTC m=+1328.359469892" Mar 13 09:34:48 crc kubenswrapper[4930]: I0313 09:34:48.546695 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:34:48 crc kubenswrapper[4930]: I0313 09:34:48.598841 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9b25b222-49ea-451d-81b8-b9fc7b54459b" containerName="glance-log" containerID="cri-o://46bb7c7db7fb23dff4d800e03e2dc775a416dd3708e10130bbdb8e2654e6a0fb" gracePeriod=30 Mar 13 09:34:48 crc kubenswrapper[4930]: I0313 09:34:48.598971 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="9b25b222-49ea-451d-81b8-b9fc7b54459b" containerName="glance-httpd" containerID="cri-o://694007dc633201b93f8eaaa97d34fc707f3c2459c2ee7752f35bf2761355d889" gracePeriod=30 Mar 13 09:34:48 crc kubenswrapper[4930]: I0313 09:34:48.698812 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:34:49 crc kubenswrapper[4930]: I0313 09:34:49.616927 4930 generic.go:334] "Generic (PLEG): container finished" podID="9b25b222-49ea-451d-81b8-b9fc7b54459b" containerID="694007dc633201b93f8eaaa97d34fc707f3c2459c2ee7752f35bf2761355d889" exitCode=0 Mar 13 09:34:49 crc kubenswrapper[4930]: I0313 09:34:49.617583 4930 generic.go:334] "Generic (PLEG): container finished" podID="9b25b222-49ea-451d-81b8-b9fc7b54459b" containerID="46bb7c7db7fb23dff4d800e03e2dc775a416dd3708e10130bbdb8e2654e6a0fb" exitCode=143 Mar 13 09:34:49 crc kubenswrapper[4930]: I0313 09:34:49.617011 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b25b222-49ea-451d-81b8-b9fc7b54459b","Type":"ContainerDied","Data":"694007dc633201b93f8eaaa97d34fc707f3c2459c2ee7752f35bf2761355d889"} Mar 13 09:34:49 crc kubenswrapper[4930]: I0313 09:34:49.617954 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b25b222-49ea-451d-81b8-b9fc7b54459b","Type":"ContainerDied","Data":"46bb7c7db7fb23dff4d800e03e2dc775a416dd3708e10130bbdb8e2654e6a0fb"} Mar 13 09:34:49 crc kubenswrapper[4930]: I0313 09:34:49.618626 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8cbd73f5-ae52-40dd-9247-ce497789fba1" containerName="glance-log" containerID="cri-o://02b6cf5f63d74d4e0767c2cf29416755be61982783c7e88f3db5e700b5de4b93" gracePeriod=30 Mar 13 09:34:49 crc kubenswrapper[4930]: I0313 09:34:49.618694 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="8cbd73f5-ae52-40dd-9247-ce497789fba1" containerName="glance-httpd" containerID="cri-o://079475b3d19765c42750282c69df7d259547fab985777833c4bb81e3106533ba" gracePeriod=30 Mar 13 09:34:50 crc kubenswrapper[4930]: I0313 09:34:50.633745 4930 generic.go:334] "Generic (PLEG): container finished" podID="8cbd73f5-ae52-40dd-9247-ce497789fba1" containerID="079475b3d19765c42750282c69df7d259547fab985777833c4bb81e3106533ba" exitCode=0 Mar 13 09:34:50 crc kubenswrapper[4930]: I0313 09:34:50.634051 4930 generic.go:334] "Generic (PLEG): container finished" podID="8cbd73f5-ae52-40dd-9247-ce497789fba1" containerID="02b6cf5f63d74d4e0767c2cf29416755be61982783c7e88f3db5e700b5de4b93" exitCode=143 Mar 13 09:34:50 crc kubenswrapper[4930]: I0313 09:34:50.633948 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8cbd73f5-ae52-40dd-9247-ce497789fba1","Type":"ContainerDied","Data":"079475b3d19765c42750282c69df7d259547fab985777833c4bb81e3106533ba"} Mar 13 09:34:50 crc kubenswrapper[4930]: I0313 09:34:50.634095 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8cbd73f5-ae52-40dd-9247-ce497789fba1","Type":"ContainerDied","Data":"02b6cf5f63d74d4e0767c2cf29416755be61982783c7e88f3db5e700b5de4b93"} Mar 13 09:34:52 crc kubenswrapper[4930]: I0313 09:34:52.224570 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:34:52 crc kubenswrapper[4930]: I0313 09:34:52.292559 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-w66lq"] Mar 13 09:34:52 crc kubenswrapper[4930]: I0313 09:34:52.292781 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerName="dnsmasq-dns" containerID="cri-o://224cc21ace1e5ad0ff39b7f7a7b1da206852e0525c50eaa2f75e871eea068c02" gracePeriod=10 Mar 13 09:34:53 crc kubenswrapper[4930]: I0313 09:34:53.251116 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.175:5353: connect: connection refused" Mar 13 09:34:53 crc kubenswrapper[4930]: I0313 09:34:53.680173 4930 generic.go:334] "Generic (PLEG): container finished" podID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerID="224cc21ace1e5ad0ff39b7f7a7b1da206852e0525c50eaa2f75e871eea068c02" exitCode=0 Mar 13 09:34:53 crc kubenswrapper[4930]: I0313 09:34:53.680469 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" event={"ID":"31a3e825-3a78-4284-a9ee-56aaa6c06c83","Type":"ContainerDied","Data":"224cc21ace1e5ad0ff39b7f7a7b1da206852e0525c50eaa2f75e871eea068c02"} Mar 13 09:34:56 crc kubenswrapper[4930]: I0313 09:34:56.912541 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:56 crc kubenswrapper[4930]: I0313 09:34:56.974013 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-credential-keys\") pod \"ba1aad44-386d-4a67-8949-7a465e65ea6e\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " Mar 13 09:34:56 crc kubenswrapper[4930]: I0313 09:34:56.974054 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-scripts\") pod \"ba1aad44-386d-4a67-8949-7a465e65ea6e\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " Mar 13 09:34:56 crc kubenswrapper[4930]: I0313 09:34:56.974078 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4txrl\" (UniqueName: \"kubernetes.io/projected/ba1aad44-386d-4a67-8949-7a465e65ea6e-kube-api-access-4txrl\") pod \"ba1aad44-386d-4a67-8949-7a465e65ea6e\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " Mar 13 09:34:56 crc kubenswrapper[4930]: I0313 09:34:56.974112 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-combined-ca-bundle\") pod \"ba1aad44-386d-4a67-8949-7a465e65ea6e\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " Mar 13 09:34:56 crc kubenswrapper[4930]: I0313 09:34:56.974202 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-config-data\") pod \"ba1aad44-386d-4a67-8949-7a465e65ea6e\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " Mar 13 09:34:56 crc kubenswrapper[4930]: I0313 09:34:56.974319 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-fernet-keys\") pod \"ba1aad44-386d-4a67-8949-7a465e65ea6e\" (UID: \"ba1aad44-386d-4a67-8949-7a465e65ea6e\") " Mar 13 09:34:56 crc kubenswrapper[4930]: I0313 09:34:56.981516 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba1aad44-386d-4a67-8949-7a465e65ea6e-kube-api-access-4txrl" (OuterVolumeSpecName: "kube-api-access-4txrl") pod "ba1aad44-386d-4a67-8949-7a465e65ea6e" (UID: "ba1aad44-386d-4a67-8949-7a465e65ea6e"). InnerVolumeSpecName "kube-api-access-4txrl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:34:56 crc kubenswrapper[4930]: I0313 09:34:56.981767 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ba1aad44-386d-4a67-8949-7a465e65ea6e" (UID: "ba1aad44-386d-4a67-8949-7a465e65ea6e"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:34:56 crc kubenswrapper[4930]: I0313 09:34:56.983597 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-scripts" (OuterVolumeSpecName: "scripts") pod "ba1aad44-386d-4a67-8949-7a465e65ea6e" (UID: "ba1aad44-386d-4a67-8949-7a465e65ea6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:34:56 crc kubenswrapper[4930]: I0313 09:34:56.985876 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ba1aad44-386d-4a67-8949-7a465e65ea6e" (UID: "ba1aad44-386d-4a67-8949-7a465e65ea6e"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:34:57 crc kubenswrapper[4930]: I0313 09:34:57.014099 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-config-data" (OuterVolumeSpecName: "config-data") pod "ba1aad44-386d-4a67-8949-7a465e65ea6e" (UID: "ba1aad44-386d-4a67-8949-7a465e65ea6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:34:57 crc kubenswrapper[4930]: I0313 09:34:57.018644 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ba1aad44-386d-4a67-8949-7a465e65ea6e" (UID: "ba1aad44-386d-4a67-8949-7a465e65ea6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:34:57 crc kubenswrapper[4930]: I0313 09:34:57.077125 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:57 crc kubenswrapper[4930]: I0313 09:34:57.077210 4930 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:57 crc kubenswrapper[4930]: I0313 09:34:57.077219 4930 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:57 crc kubenswrapper[4930]: I0313 09:34:57.077228 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:57 crc kubenswrapper[4930]: I0313 09:34:57.077237 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4txrl\" (UniqueName: \"kubernetes.io/projected/ba1aad44-386d-4a67-8949-7a465e65ea6e-kube-api-access-4txrl\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:57 crc kubenswrapper[4930]: I0313 09:34:57.077246 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ba1aad44-386d-4a67-8949-7a465e65ea6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:34:57 crc kubenswrapper[4930]: I0313 09:34:57.639521 4930 scope.go:117] "RemoveContainer" containerID="d9999ba5aeb068d6c679d90f73c0151d561e7563d00ab3a7c4d7da1e740dc633" Mar 13 09:34:57 crc kubenswrapper[4930]: I0313 09:34:57.722666 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-ftp8x" event={"ID":"ba1aad44-386d-4a67-8949-7a465e65ea6e","Type":"ContainerDied","Data":"d26896a09104b76c54ac90597c5b5fdc99bf9714f2064aef68417ba4ba430773"} Mar 13 09:34:57 crc kubenswrapper[4930]: I0313 09:34:57.722740 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d26896a09104b76c54ac90597c5b5fdc99bf9714f2064aef68417ba4ba430773" Mar 13 09:34:57 crc kubenswrapper[4930]: I0313 09:34:57.722711 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-ftp8x" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.002022 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-ftp8x"] Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.010812 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-ftp8x"] Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.106934 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-mqk65"] Mar 13 09:34:58 crc kubenswrapper[4930]: E0313 09:34:58.107531 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba1aad44-386d-4a67-8949-7a465e65ea6e" containerName="keystone-bootstrap" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.107555 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba1aad44-386d-4a67-8949-7a465e65ea6e" containerName="keystone-bootstrap" Mar 13 09:34:58 crc kubenswrapper[4930]: E0313 09:34:58.107598 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da55f1c-1e80-4231-9da7-3a9ca0923b32" containerName="init" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.107606 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da55f1c-1e80-4231-9da7-3a9ca0923b32" containerName="init" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.107892 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="1da55f1c-1e80-4231-9da7-3a9ca0923b32" containerName="init" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.107909 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba1aad44-386d-4a67-8949-7a465e65ea6e" containerName="keystone-bootstrap" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.108882 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.111759 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.116884 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.117114 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ws54h" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.117187 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.117254 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.137498 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mqk65"] Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.203121 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-fernet-keys\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.203241 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-combined-ca-bundle\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.203351 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-config-data\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.203419 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-scripts\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.203516 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vfkm8\" (UniqueName: \"kubernetes.io/projected/66e9286f-5c5e-49e5-9952-8d01782f2a09-kube-api-access-vfkm8\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.203549 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-credential-keys\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.249518 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.175:5353: connect: connection refused" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.307328 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-fernet-keys\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.307413 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-combined-ca-bundle\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.307473 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-config-data\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.307511 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-scripts\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.307564 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vfkm8\" (UniqueName: \"kubernetes.io/projected/66e9286f-5c5e-49e5-9952-8d01782f2a09-kube-api-access-vfkm8\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.307584 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-credential-keys\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.313062 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-config-data\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.313176 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-combined-ca-bundle\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.314921 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-credential-keys\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.315021 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-fernet-keys\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.322351 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-scripts\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.323005 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vfkm8\" (UniqueName: \"kubernetes.io/projected/66e9286f-5c5e-49e5-9952-8d01782f2a09-kube-api-access-vfkm8\") pod \"keystone-bootstrap-mqk65\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:58 crc kubenswrapper[4930]: I0313 09:34:58.433186 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:34:59 crc kubenswrapper[4930]: I0313 09:34:59.985770 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba1aad44-386d-4a67-8949-7a465e65ea6e" path="/var/lib/kubelet/pods/ba1aad44-386d-4a67-8949-7a465e65ea6e/volumes" Mar 13 09:35:00 crc kubenswrapper[4930]: E0313 09:35:00.109353 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Mar 13 09:35:00 crc kubenswrapper[4930]: E0313 09:35:00.109587 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4pmwb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-h9bf6_openstack(d3fb676a-c7df-4e03-a4ad-41da78bde051): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:35:00 crc kubenswrapper[4930]: E0313 09:35:00.112271 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-h9bf6" podUID="d3fb676a-c7df-4e03-a4ad-41da78bde051" Mar 13 09:35:00 crc kubenswrapper[4930]: E0313 09:35:00.753393 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-h9bf6" podUID="d3fb676a-c7df-4e03-a4ad-41da78bde051" Mar 13 09:35:01 crc kubenswrapper[4930]: E0313 09:35:01.594062 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Mar 13 09:35:01 crc kubenswrapper[4930]: E0313 09:35:01.594818 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bqwxf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-djnqj_openstack(10e08b2d-bcbf-4b39-9baa-e4ca5213567c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:35:01 crc kubenswrapper[4930]: E0313 09:35:01.596019 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-djnqj" podUID="10e08b2d-bcbf-4b39-9baa-e4ca5213567c" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.700768 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.765159 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.765344 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"8cbd73f5-ae52-40dd-9247-ce497789fba1","Type":"ContainerDied","Data":"3e579889c31aac96e4cb18129fd4cdfb7e39a6b2d4ea24140129d2f4393c104a"} Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.765456 4930 scope.go:117] "RemoveContainer" containerID="079475b3d19765c42750282c69df7d259547fab985777833c4bb81e3106533ba" Mar 13 09:35:01 crc kubenswrapper[4930]: E0313 09:35:01.767703 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-djnqj" podUID="10e08b2d-bcbf-4b39-9baa-e4ca5213567c" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.791279 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8cbd73f5-ae52-40dd-9247-ce497789fba1-httpd-run\") pod \"8cbd73f5-ae52-40dd-9247-ce497789fba1\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.791605 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtssp\" (UniqueName: \"kubernetes.io/projected/8cbd73f5-ae52-40dd-9247-ce497789fba1-kube-api-access-gtssp\") pod \"8cbd73f5-ae52-40dd-9247-ce497789fba1\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.791716 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-config-data\") pod \"8cbd73f5-ae52-40dd-9247-ce497789fba1\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.791742 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-combined-ca-bundle\") pod \"8cbd73f5-ae52-40dd-9247-ce497789fba1\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.791810 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cbd73f5-ae52-40dd-9247-ce497789fba1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8cbd73f5-ae52-40dd-9247-ce497789fba1" (UID: "8cbd73f5-ae52-40dd-9247-ce497789fba1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.791830 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbd73f5-ae52-40dd-9247-ce497789fba1-logs\") pod \"8cbd73f5-ae52-40dd-9247-ce497789fba1\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.791856 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-scripts\") pod \"8cbd73f5-ae52-40dd-9247-ce497789fba1\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.791956 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"8cbd73f5-ae52-40dd-9247-ce497789fba1\" (UID: \"8cbd73f5-ae52-40dd-9247-ce497789fba1\") " Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.792633 4930 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8cbd73f5-ae52-40dd-9247-ce497789fba1-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.792804 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8cbd73f5-ae52-40dd-9247-ce497789fba1-logs" (OuterVolumeSpecName: "logs") pod "8cbd73f5-ae52-40dd-9247-ce497789fba1" (UID: "8cbd73f5-ae52-40dd-9247-ce497789fba1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.805018 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-scripts" (OuterVolumeSpecName: "scripts") pod "8cbd73f5-ae52-40dd-9247-ce497789fba1" (UID: "8cbd73f5-ae52-40dd-9247-ce497789fba1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.805066 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cbd73f5-ae52-40dd-9247-ce497789fba1-kube-api-access-gtssp" (OuterVolumeSpecName: "kube-api-access-gtssp") pod "8cbd73f5-ae52-40dd-9247-ce497789fba1" (UID: "8cbd73f5-ae52-40dd-9247-ce497789fba1"). InnerVolumeSpecName "kube-api-access-gtssp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.820279 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f" (OuterVolumeSpecName: "glance") pod "8cbd73f5-ae52-40dd-9247-ce497789fba1" (UID: "8cbd73f5-ae52-40dd-9247-ce497789fba1"). InnerVolumeSpecName "pvc-893414a0-cbec-42bd-a89f-5c787585718f". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.838178 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8cbd73f5-ae52-40dd-9247-ce497789fba1" (UID: "8cbd73f5-ae52-40dd-9247-ce497789fba1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.859634 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-config-data" (OuterVolumeSpecName: "config-data") pod "8cbd73f5-ae52-40dd-9247-ce497789fba1" (UID: "8cbd73f5-ae52-40dd-9247-ce497789fba1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.894805 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtssp\" (UniqueName: \"kubernetes.io/projected/8cbd73f5-ae52-40dd-9247-ce497789fba1-kube-api-access-gtssp\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.894837 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.894848 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.894856 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8cbd73f5-ae52-40dd-9247-ce497789fba1-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.894864 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8cbd73f5-ae52-40dd-9247-ce497789fba1-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.894888 4930 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") on node \"crc\" " Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.920414 4930 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.922041 4930 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-893414a0-cbec-42bd-a89f-5c787585718f" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f") on node "crc" Mar 13 09:35:01 crc kubenswrapper[4930]: I0313 09:35:01.996616 4930 reconciler_common.go:293] "Volume detached for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:02 crc kubenswrapper[4930]: E0313 09:35:02.071596 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Mar 13 09:35:02 crc kubenswrapper[4930]: E0313 09:35:02.071785 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-d9m4t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-p4nk9_openstack(ae61cc80-b8f2-401c-87db-5728d566b288): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:35:02 crc kubenswrapper[4930]: E0313 09:35:02.073587 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-p4nk9" podUID="ae61cc80-b8f2-401c-87db-5728d566b288" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.103499 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.115898 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.132507 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:35:02 crc kubenswrapper[4930]: E0313 09:35:02.133085 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbd73f5-ae52-40dd-9247-ce497789fba1" containerName="glance-log" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.133105 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbd73f5-ae52-40dd-9247-ce497789fba1" containerName="glance-log" Mar 13 09:35:02 crc kubenswrapper[4930]: E0313 09:35:02.133121 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cbd73f5-ae52-40dd-9247-ce497789fba1" containerName="glance-httpd" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.133128 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cbd73f5-ae52-40dd-9247-ce497789fba1" containerName="glance-httpd" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.133315 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cbd73f5-ae52-40dd-9247-ce497789fba1" containerName="glance-log" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.133339 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cbd73f5-ae52-40dd-9247-ce497789fba1" containerName="glance-httpd" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.134480 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.136572 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.136798 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.144903 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.208396 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2srw\" (UniqueName: \"kubernetes.io/projected/178d7163-e4cc-4d25-9f42-2182928d43e4-kube-api-access-g2srw\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.208482 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.208572 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/178d7163-e4cc-4d25-9f42-2182928d43e4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.208657 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/178d7163-e4cc-4d25-9f42-2182928d43e4-logs\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.208698 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.208731 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.208764 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.208920 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.311128 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/178d7163-e4cc-4d25-9f42-2182928d43e4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.311211 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/178d7163-e4cc-4d25-9f42-2182928d43e4-logs\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.311236 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.311255 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.311278 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.311353 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.311418 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2srw\" (UniqueName: \"kubernetes.io/projected/178d7163-e4cc-4d25-9f42-2182928d43e4-kube-api-access-g2srw\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.311461 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.312991 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/178d7163-e4cc-4d25-9f42-2182928d43e4-logs\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.313215 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/178d7163-e4cc-4d25-9f42-2182928d43e4-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.315762 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.315982 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-scripts\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.317344 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-config-data\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.317724 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.319609 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.319867 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a4f4242b5f3665838ee27f4f8b1e1823b22b221bc7d207651abbb8d37832373d/globalmount\"" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.333081 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2srw\" (UniqueName: \"kubernetes.io/projected/178d7163-e4cc-4d25-9f42-2182928d43e4-kube-api-access-g2srw\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.359172 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"glance-default-internal-api-0\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: I0313 09:35:02.474609 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:02 crc kubenswrapper[4930]: E0313 09:35:02.776859 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-p4nk9" podUID="ae61cc80-b8f2-401c-87db-5728d566b288" Mar 13 09:35:03 crc kubenswrapper[4930]: I0313 09:35:03.984897 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cbd73f5-ae52-40dd-9247-ce497789fba1" path="/var/lib/kubelet/pods/8cbd73f5-ae52-40dd-9247-ce497789fba1/volumes" Mar 13 09:35:06 crc kubenswrapper[4930]: I0313 09:35:06.831300 4930 generic.go:334] "Generic (PLEG): container finished" podID="d37e8e55-e1f6-4c48-8f14-e4d714afb1fc" containerID="681d0cd0f264d27cd661797cc1039f195824ca91fbb3eac1e6d8422a9da38dc1" exitCode=0 Mar 13 09:35:06 crc kubenswrapper[4930]: I0313 09:35:06.831384 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-pdp9j" event={"ID":"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc","Type":"ContainerDied","Data":"681d0cd0f264d27cd661797cc1039f195824ca91fbb3eac1e6d8422a9da38dc1"} Mar 13 09:35:08 crc kubenswrapper[4930]: I0313 09:35:08.257510 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.175:5353: i/o timeout" Mar 13 09:35:08 crc kubenswrapper[4930]: I0313 09:35:08.258179 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:35:12 crc kubenswrapper[4930]: I0313 09:35:12.308762 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:35:12 crc kubenswrapper[4930]: I0313 09:35:12.309295 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:35:12 crc kubenswrapper[4930]: I0313 09:35:12.309355 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:35:12 crc kubenswrapper[4930]: I0313 09:35:12.310136 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9a600e0833fe6da30e3cb3376c93b675518bc4f00e36348b0e36bfb8a45c09d7"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 09:35:12 crc kubenswrapper[4930]: I0313 09:35:12.310261 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://9a600e0833fe6da30e3cb3376c93b675518bc4f00e36348b0e36bfb8a45c09d7" gracePeriod=600 Mar 13 09:35:12 crc kubenswrapper[4930]: I0313 09:35:12.969515 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 09:35:12 crc kubenswrapper[4930]: I0313 09:35:12.969566 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 09:35:13 crc kubenswrapper[4930]: I0313 09:35:13.259065 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.175:5353: i/o timeout" Mar 13 09:35:14 crc kubenswrapper[4930]: I0313 09:35:14.288489 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 09:35:14 crc kubenswrapper[4930]: I0313 09:35:14.341778 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="9a600e0833fe6da30e3cb3376c93b675518bc4f00e36348b0e36bfb8a45c09d7" exitCode=0 Mar 13 09:35:14 crc kubenswrapper[4930]: I0313 09:35:14.341884 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"9a600e0833fe6da30e3cb3376c93b675518bc4f00e36348b0e36bfb8a45c09d7"} Mar 13 09:35:18 crc kubenswrapper[4930]: I0313 09:35:18.260550 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.175:5353: i/o timeout" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.520115 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.527729 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.534900 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.644687 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" event={"ID":"31a3e825-3a78-4284-a9ee-56aaa6c06c83","Type":"ContainerDied","Data":"0a5ff13ada066aa843658cef3bc8aaa5d218ebf5d3ce0d7f549fc0e4e90a5096"} Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.644728 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.646616 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-pdp9j" event={"ID":"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc","Type":"ContainerDied","Data":"4c5af5cec93f7f8bc461701a1335dac3cef8fe75b0f13dba8493cc37bc39c6bd"} Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.646649 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c5af5cec93f7f8bc461701a1335dac3cef8fe75b0f13dba8493cc37bc39c6bd" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.646629 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-pdp9j" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.648849 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"9b25b222-49ea-451d-81b8-b9fc7b54459b","Type":"ContainerDied","Data":"e0a97ab608d97e88728aa48b517ffcc138e8ca7b991f888a0ae746688d02f3ed"} Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.648905 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679239 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b25b222-49ea-451d-81b8-b9fc7b54459b-logs\") pod \"9b25b222-49ea-451d-81b8-b9fc7b54459b\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679306 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-combined-ca-bundle\") pod \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\" (UID: \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679353 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b25b222-49ea-451d-81b8-b9fc7b54459b-httpd-run\") pod \"9b25b222-49ea-451d-81b8-b9fc7b54459b\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679418 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-ovsdbserver-nb\") pod \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679472 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-combined-ca-bundle\") pod \"9b25b222-49ea-451d-81b8-b9fc7b54459b\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679512 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-config-data\") pod \"9b25b222-49ea-451d-81b8-b9fc7b54459b\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679545 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-scripts\") pod \"9b25b222-49ea-451d-81b8-b9fc7b54459b\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679580 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-dns-swift-storage-0\") pod \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679664 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wssn4\" (UniqueName: \"kubernetes.io/projected/9b25b222-49ea-451d-81b8-b9fc7b54459b-kube-api-access-wssn4\") pod \"9b25b222-49ea-451d-81b8-b9fc7b54459b\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679749 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-config\") pod \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679789 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-config\") pod \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\" (UID: \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679821 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-dns-svc\") pod \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679846 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whtfb\" (UniqueName: \"kubernetes.io/projected/31a3e825-3a78-4284-a9ee-56aaa6c06c83-kube-api-access-whtfb\") pod \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679984 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"9b25b222-49ea-451d-81b8-b9fc7b54459b\" (UID: \"9b25b222-49ea-451d-81b8-b9fc7b54459b\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.680015 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-ovsdbserver-sb\") pod \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\" (UID: \"31a3e825-3a78-4284-a9ee-56aaa6c06c83\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.680068 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf2b4\" (UniqueName: \"kubernetes.io/projected/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-kube-api-access-hf2b4\") pod \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\" (UID: \"d37e8e55-e1f6-4c48-8f14-e4d714afb1fc\") " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679834 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b25b222-49ea-451d-81b8-b9fc7b54459b-logs" (OuterVolumeSpecName: "logs") pod "9b25b222-49ea-451d-81b8-b9fc7b54459b" (UID: "9b25b222-49ea-451d-81b8-b9fc7b54459b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.679865 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b25b222-49ea-451d-81b8-b9fc7b54459b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9b25b222-49ea-451d-81b8-b9fc7b54459b" (UID: "9b25b222-49ea-451d-81b8-b9fc7b54459b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.685595 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-kube-api-access-hf2b4" (OuterVolumeSpecName: "kube-api-access-hf2b4") pod "d37e8e55-e1f6-4c48-8f14-e4d714afb1fc" (UID: "d37e8e55-e1f6-4c48-8f14-e4d714afb1fc"). InnerVolumeSpecName "kube-api-access-hf2b4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.686285 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31a3e825-3a78-4284-a9ee-56aaa6c06c83-kube-api-access-whtfb" (OuterVolumeSpecName: "kube-api-access-whtfb") pod "31a3e825-3a78-4284-a9ee-56aaa6c06c83" (UID: "31a3e825-3a78-4284-a9ee-56aaa6c06c83"). InnerVolumeSpecName "kube-api-access-whtfb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.686373 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf2b4\" (UniqueName: \"kubernetes.io/projected/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-kube-api-access-hf2b4\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.686395 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9b25b222-49ea-451d-81b8-b9fc7b54459b-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.686409 4930 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9b25b222-49ea-451d-81b8-b9fc7b54459b-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.697957 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-scripts" (OuterVolumeSpecName: "scripts") pod "9b25b222-49ea-451d-81b8-b9fc7b54459b" (UID: "9b25b222-49ea-451d-81b8-b9fc7b54459b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.705588 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958" (OuterVolumeSpecName: "glance") pod "9b25b222-49ea-451d-81b8-b9fc7b54459b" (UID: "9b25b222-49ea-451d-81b8-b9fc7b54459b"). InnerVolumeSpecName "pvc-e6a7410f-e5ed-48ef-993c-0deb96840958". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.714677 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b25b222-49ea-451d-81b8-b9fc7b54459b-kube-api-access-wssn4" (OuterVolumeSpecName: "kube-api-access-wssn4") pod "9b25b222-49ea-451d-81b8-b9fc7b54459b" (UID: "9b25b222-49ea-451d-81b8-b9fc7b54459b"). InnerVolumeSpecName "kube-api-access-wssn4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.736936 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "31a3e825-3a78-4284-a9ee-56aaa6c06c83" (UID: "31a3e825-3a78-4284-a9ee-56aaa6c06c83"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.750471 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d37e8e55-e1f6-4c48-8f14-e4d714afb1fc" (UID: "d37e8e55-e1f6-4c48-8f14-e4d714afb1fc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.752616 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-config" (OuterVolumeSpecName: "config") pod "d37e8e55-e1f6-4c48-8f14-e4d714afb1fc" (UID: "d37e8e55-e1f6-4c48-8f14-e4d714afb1fc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.753847 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "31a3e825-3a78-4284-a9ee-56aaa6c06c83" (UID: "31a3e825-3a78-4284-a9ee-56aaa6c06c83"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.760442 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "31a3e825-3a78-4284-a9ee-56aaa6c06c83" (UID: "31a3e825-3a78-4284-a9ee-56aaa6c06c83"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.764865 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9b25b222-49ea-451d-81b8-b9fc7b54459b" (UID: "9b25b222-49ea-451d-81b8-b9fc7b54459b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.770828 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-config-data" (OuterVolumeSpecName: "config-data") pod "9b25b222-49ea-451d-81b8-b9fc7b54459b" (UID: "9b25b222-49ea-451d-81b8-b9fc7b54459b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.773743 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "31a3e825-3a78-4284-a9ee-56aaa6c06c83" (UID: "31a3e825-3a78-4284-a9ee-56aaa6c06c83"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.787783 4930 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.787805 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wssn4\" (UniqueName: \"kubernetes.io/projected/9b25b222-49ea-451d-81b8-b9fc7b54459b-kube-api-access-wssn4\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.787820 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.787832 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.787845 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whtfb\" (UniqueName: \"kubernetes.io/projected/31a3e825-3a78-4284-a9ee-56aaa6c06c83-kube-api-access-whtfb\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.787931 4930 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") on node \"crc\" " Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.787950 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.787963 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.787974 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.787984 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.787993 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.788001 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9b25b222-49ea-451d-81b8-b9fc7b54459b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.803189 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-config" (OuterVolumeSpecName: "config") pod "31a3e825-3a78-4284-a9ee-56aaa6c06c83" (UID: "31a3e825-3a78-4284-a9ee-56aaa6c06c83"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.820632 4930 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.820773 4930 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e6a7410f-e5ed-48ef-993c-0deb96840958" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958") on node "crc" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.889657 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31a3e825-3a78-4284-a9ee-56aaa6c06c83-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.889693 4930 reconciler_common.go:293] "Volume detached for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:20 crc kubenswrapper[4930]: I0313 09:35:20.987229 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-w66lq"] Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.000200 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c79d794d7-w66lq"] Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.011180 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.029572 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.052266 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:35:21 crc kubenswrapper[4930]: E0313 09:35:21.052798 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d37e8e55-e1f6-4c48-8f14-e4d714afb1fc" containerName="neutron-db-sync" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.052824 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d37e8e55-e1f6-4c48-8f14-e4d714afb1fc" containerName="neutron-db-sync" Mar 13 09:35:21 crc kubenswrapper[4930]: E0313 09:35:21.052858 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerName="dnsmasq-dns" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.052868 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerName="dnsmasq-dns" Mar 13 09:35:21 crc kubenswrapper[4930]: E0313 09:35:21.052883 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b25b222-49ea-451d-81b8-b9fc7b54459b" containerName="glance-httpd" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.052891 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b25b222-49ea-451d-81b8-b9fc7b54459b" containerName="glance-httpd" Mar 13 09:35:21 crc kubenswrapper[4930]: E0313 09:35:21.052915 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerName="init" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.052924 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerName="init" Mar 13 09:35:21 crc kubenswrapper[4930]: E0313 09:35:21.052938 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b25b222-49ea-451d-81b8-b9fc7b54459b" containerName="glance-log" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.052944 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b25b222-49ea-451d-81b8-b9fc7b54459b" containerName="glance-log" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.053169 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d37e8e55-e1f6-4c48-8f14-e4d714afb1fc" containerName="neutron-db-sync" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.053183 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerName="dnsmasq-dns" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.053210 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b25b222-49ea-451d-81b8-b9fc7b54459b" containerName="glance-httpd" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.053221 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b25b222-49ea-451d-81b8-b9fc7b54459b" containerName="glance-log" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.054353 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.058472 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.062141 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.085774 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.208467 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nrpc\" (UniqueName: \"kubernetes.io/projected/d95152a3-5e16-464b-9785-88e629cc31f9-kube-api-access-8nrpc\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.208533 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.208567 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d95152a3-5e16-464b-9785-88e629cc31f9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.208644 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.208698 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-scripts\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.208753 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-config-data\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.208788 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d95152a3-5e16-464b-9785-88e629cc31f9-logs\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.208822 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.312653 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-scripts\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.312784 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-config-data\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.312838 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d95152a3-5e16-464b-9785-88e629cc31f9-logs\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.312891 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.313003 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nrpc\" (UniqueName: \"kubernetes.io/projected/d95152a3-5e16-464b-9785-88e629cc31f9-kube-api-access-8nrpc\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.313046 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.313083 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d95152a3-5e16-464b-9785-88e629cc31f9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.313191 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.313326 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d95152a3-5e16-464b-9785-88e629cc31f9-logs\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.314122 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d95152a3-5e16-464b-9785-88e629cc31f9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.318012 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-config-data\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.318983 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.319164 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-scripts\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.319454 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.319492 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.319520 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/99223c607948bed00b9fe5109399df143422767dfa8a32b71d4b371c3748c0ce/globalmount\"" pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.334358 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nrpc\" (UniqueName: \"kubernetes.io/projected/d95152a3-5e16-464b-9785-88e629cc31f9-kube-api-access-8nrpc\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.368713 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"glance-default-external-api-0\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.374391 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.746758 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-5hc9m"] Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.754949 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.778233 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-5hc9m"] Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.832234 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-dns-svc\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.832508 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.832573 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-config\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.832684 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.832931 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.833071 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kw6j\" (UniqueName: \"kubernetes.io/projected/9a434276-b75a-465b-b5d8-d75dc7624d09-kube-api-access-2kw6j\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.873925 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6dcf4d8cb8-7d6r6"] Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.878556 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.883315 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.883635 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.883767 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-l72lb" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.884860 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.896788 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6dcf4d8cb8-7d6r6"] Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.935265 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-dns-svc\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.935363 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.935397 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-config\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.935443 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.935495 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.935517 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kw6j\" (UniqueName: \"kubernetes.io/projected/9a434276-b75a-465b-b5d8-d75dc7624d09-kube-api-access-2kw6j\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.936511 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-dns-svc\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.936538 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-config\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.937358 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.938306 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.940698 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.956853 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kw6j\" (UniqueName: \"kubernetes.io/projected/9a434276-b75a-465b-b5d8-d75dc7624d09-kube-api-access-2kw6j\") pod \"dnsmasq-dns-6b7b667979-5hc9m\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.994091 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" path="/var/lib/kubelet/pods/31a3e825-3a78-4284-a9ee-56aaa6c06c83/volumes" Mar 13 09:35:21 crc kubenswrapper[4930]: I0313 09:35:21.996095 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b25b222-49ea-451d-81b8-b9fc7b54459b" path="/var/lib/kubelet/pods/9b25b222-49ea-451d-81b8-b9fc7b54459b/volumes" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.037340 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-ovndb-tls-certs\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.037421 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvq4c\" (UniqueName: \"kubernetes.io/projected/aef74c6a-403d-4388-a873-43dca210f617-kube-api-access-tvq4c\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.037515 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-config\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.037554 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-httpd-config\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.037579 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-combined-ca-bundle\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.129850 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.140368 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-ovndb-tls-certs\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.140450 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvq4c\" (UniqueName: \"kubernetes.io/projected/aef74c6a-403d-4388-a873-43dca210f617-kube-api-access-tvq4c\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.140508 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-config\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.140551 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-httpd-config\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.140579 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-combined-ca-bundle\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.144837 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-config\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.148584 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-ovndb-tls-certs\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.148730 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-combined-ca-bundle\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: E0313 09:35:22.152741 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Mar 13 09:35:22 crc kubenswrapper[4930]: E0313 09:35:22.152969 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gmcpt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-7dsrx_openstack(98c022d1-9a39-4f3e-8108-e4be2b287077): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.153292 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-httpd-config\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: E0313 09:35:22.154564 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-7dsrx" podUID="98c022d1-9a39-4f3e-8108-e4be2b287077" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.167291 4930 scope.go:117] "RemoveContainer" containerID="02b6cf5f63d74d4e0767c2cf29416755be61982783c7e88f3db5e700b5de4b93" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.179094 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvq4c\" (UniqueName: \"kubernetes.io/projected/aef74c6a-403d-4388-a873-43dca210f617-kube-api-access-tvq4c\") pod \"neutron-6dcf4d8cb8-7d6r6\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.218086 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.452588 4930 scope.go:117] "RemoveContainer" containerID="c06f278e05cf4208d0a63eab8c9e20f18523a288ed0cb0082dfaac4bd5791ab7" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.743503 4930 scope.go:117] "RemoveContainer" containerID="224cc21ace1e5ad0ff39b7f7a7b1da206852e0525c50eaa2f75e871eea068c02" Mar 13 09:35:22 crc kubenswrapper[4930]: E0313 09:35:22.744658 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-7dsrx" podUID="98c022d1-9a39-4f3e-8108-e4be2b287077" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.778032 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-mqk65"] Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.848883 4930 scope.go:117] "RemoveContainer" containerID="5460f65f3e2b8821d1a4bdb4576a88d3abd7232ee3465848f2d08959151b5b0d" Mar 13 09:35:22 crc kubenswrapper[4930]: I0313 09:35:22.986897 4930 scope.go:117] "RemoveContainer" containerID="694007dc633201b93f8eaaa97d34fc707f3c2459c2ee7752f35bf2761355d889" Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.067337 4930 scope.go:117] "RemoveContainer" containerID="46bb7c7db7fb23dff4d800e03e2dc775a416dd3708e10130bbdb8e2654e6a0fb" Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.253473 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.261581 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5c79d794d7-w66lq" podUID="31a3e825-3a78-4284-a9ee-56aaa6c06c83" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.175:5353: i/o timeout" Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.376521 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.385966 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-5hc9m"] Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.484237 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6dcf4d8cb8-7d6r6"] Mar 13 09:35:23 crc kubenswrapper[4930]: W0313 09:35:23.500326 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a434276_b75a_465b_b5d8_d75dc7624d09.slice/crio-1bcfa8ff35ca2fdbd05533fa07cf290e83c83b3f1b3a0f4636fc2a3c11cef7dc WatchSource:0}: Error finding container 1bcfa8ff35ca2fdbd05533fa07cf290e83c83b3f1b3a0f4636fc2a3c11cef7dc: Status 404 returned error can't find the container with id 1bcfa8ff35ca2fdbd05533fa07cf290e83c83b3f1b3a0f4636fc2a3c11cef7dc Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.743922 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc498063-054b-4708-981d-65da056e5b2f","Type":"ContainerStarted","Data":"02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff"} Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.746588 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-djnqj" event={"ID":"10e08b2d-bcbf-4b39-9baa-e4ca5213567c","Type":"ContainerStarted","Data":"cb71701e4d8a02633aae086604550da618517bad71a9e78fdb788997e04e82c5"} Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.759031 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" event={"ID":"9a434276-b75a-465b-b5d8-d75dc7624d09","Type":"ContainerStarted","Data":"1bcfa8ff35ca2fdbd05533fa07cf290e83c83b3f1b3a0f4636fc2a3c11cef7dc"} Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.761717 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dcf4d8cb8-7d6r6" event={"ID":"aef74c6a-403d-4388-a873-43dca210f617","Type":"ContainerStarted","Data":"daf06e22e6999efd5421ff9c6a42745bf71206ffd67ff1c23619fd861fb3d33a"} Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.792479 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqk65" event={"ID":"66e9286f-5c5e-49e5-9952-8d01782f2a09","Type":"ContainerStarted","Data":"c959109492edd2baef172e975037522c981f72caba726ee20c38339cc13fd18d"} Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.792534 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqk65" event={"ID":"66e9286f-5c5e-49e5-9952-8d01782f2a09","Type":"ContainerStarted","Data":"b36f7adb10d8edcf64730260c9c0706254531d64b1a252ad91843b80116c2e4e"} Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.804314 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-djnqj" podStartSLOduration=3.970506544 podStartE2EDuration="45.804292936s" podCreationTimestamp="2026-03-13 09:34:38 +0000 UTC" firstStartedPulling="2026-03-13 09:34:40.618575646 +0000 UTC m=+1321.368490323" lastFinishedPulling="2026-03-13 09:35:22.452362038 +0000 UTC m=+1363.202276715" observedRunningTime="2026-03-13 09:35:23.769707481 +0000 UTC m=+1364.519622158" watchObservedRunningTime="2026-03-13 09:35:23.804292936 +0000 UTC m=+1364.554207613" Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.824235 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766"} Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.837802 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-mqk65" podStartSLOduration=25.837779303 podStartE2EDuration="25.837779303s" podCreationTimestamp="2026-03-13 09:34:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:23.828656364 +0000 UTC m=+1364.578571061" watchObservedRunningTime="2026-03-13 09:35:23.837779303 +0000 UTC m=+1364.587693980" Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.876196 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"178d7163-e4cc-4d25-9f42-2182928d43e4","Type":"ContainerStarted","Data":"941de66ab894a71ec706670847807fc7ee1a23b1f03cc5a89f3b17afb5224c04"} Mar 13 09:35:23 crc kubenswrapper[4930]: I0313 09:35:23.877486 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d95152a3-5e16-464b-9785-88e629cc31f9","Type":"ContainerStarted","Data":"a54fd8e66dc7f4c2b59d1ec89408b71f3a5fdfaa5eac229d6321b224230d254e"} Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.502170 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-75f4c75545-5xgvn"] Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.504399 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.507609 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.514773 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.518560 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-75f4c75545-5xgvn"] Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.559153 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-ovndb-tls-certs\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.559230 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-httpd-config\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.559271 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-config\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.559362 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-internal-tls-certs\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.559381 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-public-tls-certs\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.559415 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnks8\" (UniqueName: \"kubernetes.io/projected/46162d09-ed14-4445-9644-3658980a85e5-kube-api-access-rnks8\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.559436 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-combined-ca-bundle\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.662722 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-httpd-config\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.663299 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-config\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.664837 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-internal-tls-certs\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.664918 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-public-tls-certs\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.665053 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnks8\" (UniqueName: \"kubernetes.io/projected/46162d09-ed14-4445-9644-3658980a85e5-kube-api-access-rnks8\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.665138 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-combined-ca-bundle\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.665401 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-ovndb-tls-certs\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.667807 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-config\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.671614 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-httpd-config\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.672137 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-ovndb-tls-certs\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.674117 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-combined-ca-bundle\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.674142 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-public-tls-certs\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.674532 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-internal-tls-certs\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.698136 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnks8\" (UniqueName: \"kubernetes.io/projected/46162d09-ed14-4445-9644-3658980a85e5-kube-api-access-rnks8\") pod \"neutron-75f4c75545-5xgvn\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.832039 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.903009 4930 generic.go:334] "Generic (PLEG): container finished" podID="9a434276-b75a-465b-b5d8-d75dc7624d09" containerID="2f4c8976574acf8e6b97bfb6522062d106f8fa43e8bd739b4e739aaef12953ec" exitCode=0 Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.903071 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" event={"ID":"9a434276-b75a-465b-b5d8-d75dc7624d09","Type":"ContainerDied","Data":"2f4c8976574acf8e6b97bfb6522062d106f8fa43e8bd739b4e739aaef12953ec"} Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.917598 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d95152a3-5e16-464b-9785-88e629cc31f9","Type":"ContainerStarted","Data":"995636287c1ceecdc41a41988d0a3367c860c034253f4dffe261994bdfc30a8f"} Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.921245 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-p4nk9" event={"ID":"ae61cc80-b8f2-401c-87db-5728d566b288","Type":"ContainerStarted","Data":"6e35cb5389d90f31f41ef51b7c6a85b9edc8e6f4e042fbf370a1bba9fd1433f9"} Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.953098 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-p4nk9" podStartSLOduration=5.035879932 podStartE2EDuration="46.953080298s" podCreationTimestamp="2026-03-13 09:34:38 +0000 UTC" firstStartedPulling="2026-03-13 09:34:40.58665896 +0000 UTC m=+1321.336573637" lastFinishedPulling="2026-03-13 09:35:22.503859326 +0000 UTC m=+1363.253774003" observedRunningTime="2026-03-13 09:35:24.94779669 +0000 UTC m=+1365.697711377" watchObservedRunningTime="2026-03-13 09:35:24.953080298 +0000 UTC m=+1365.702994975" Mar 13 09:35:24 crc kubenswrapper[4930]: I0313 09:35:24.980764 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h9bf6" event={"ID":"d3fb676a-c7df-4e03-a4ad-41da78bde051","Type":"ContainerStarted","Data":"ac95d58ba906ac3a4c40acd96be960d3b3dfc8dacf73f640c3c779d6f7b101df"} Mar 13 09:35:25 crc kubenswrapper[4930]: I0313 09:35:25.013054 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-h9bf6" podStartSLOduration=4.654201081 podStartE2EDuration="47.013029177s" podCreationTimestamp="2026-03-13 09:34:38 +0000 UTC" firstStartedPulling="2026-03-13 09:34:40.114019208 +0000 UTC m=+1320.863933885" lastFinishedPulling="2026-03-13 09:35:22.472847304 +0000 UTC m=+1363.222761981" observedRunningTime="2026-03-13 09:35:25.005600773 +0000 UTC m=+1365.755515460" watchObservedRunningTime="2026-03-13 09:35:25.013029177 +0000 UTC m=+1365.762943864" Mar 13 09:35:25 crc kubenswrapper[4930]: I0313 09:35:25.027990 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dcf4d8cb8-7d6r6" event={"ID":"aef74c6a-403d-4388-a873-43dca210f617","Type":"ContainerStarted","Data":"35ba88787bfb2c888f48f45e752c3ba2753ebbdbab7fbe30723229f6d66e42fc"} Mar 13 09:35:25 crc kubenswrapper[4930]: I0313 09:35:25.028036 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dcf4d8cb8-7d6r6" event={"ID":"aef74c6a-403d-4388-a873-43dca210f617","Type":"ContainerStarted","Data":"93155ac78aa8e7896ad1c995dc5aac22c196b3c9ecdcd86f67cbd7bfb454b789"} Mar 13 09:35:25 crc kubenswrapper[4930]: I0313 09:35:25.028251 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:25 crc kubenswrapper[4930]: I0313 09:35:25.069036 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"178d7163-e4cc-4d25-9f42-2182928d43e4","Type":"ContainerStarted","Data":"b4d57afb3f9f4b4daa1177a0a7464e8a09b60d23e099df13c0deae7d6fbc4319"} Mar 13 09:35:25 crc kubenswrapper[4930]: I0313 09:35:25.096669 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6dcf4d8cb8-7d6r6" podStartSLOduration=4.096641736 podStartE2EDuration="4.096641736s" podCreationTimestamp="2026-03-13 09:35:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:25.065402638 +0000 UTC m=+1365.815317325" watchObservedRunningTime="2026-03-13 09:35:25.096641736 +0000 UTC m=+1365.846556423" Mar 13 09:35:25 crc kubenswrapper[4930]: I0313 09:35:25.559859 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-75f4c75545-5xgvn"] Mar 13 09:35:26 crc kubenswrapper[4930]: I0313 09:35:26.079208 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d95152a3-5e16-464b-9785-88e629cc31f9","Type":"ContainerStarted","Data":"29fb594a8475f74faf9b69578f86c18f499f27eb837df8c5196f92b52408d168"} Mar 13 09:35:26 crc kubenswrapper[4930]: I0313 09:35:26.082768 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"178d7163-e4cc-4d25-9f42-2182928d43e4","Type":"ContainerStarted","Data":"c6936094e864674206cc13c48094456a401ee0af5803bd17c934cdc8e261e538"} Mar 13 09:35:26 crc kubenswrapper[4930]: I0313 09:35:26.195178 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.195152921 podStartE2EDuration="5.195152921s" podCreationTimestamp="2026-03-13 09:35:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:26.157913807 +0000 UTC m=+1366.907828484" watchObservedRunningTime="2026-03-13 09:35:26.195152921 +0000 UTC m=+1366.945067598" Mar 13 09:35:26 crc kubenswrapper[4930]: I0313 09:35:26.199912 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=24.199891525 podStartE2EDuration="24.199891525s" podCreationTimestamp="2026-03-13 09:35:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:26.189796771 +0000 UTC m=+1366.939711448" watchObservedRunningTime="2026-03-13 09:35:26.199891525 +0000 UTC m=+1366.949806202" Mar 13 09:35:27 crc kubenswrapper[4930]: I0313 09:35:27.102856 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" event={"ID":"9a434276-b75a-465b-b5d8-d75dc7624d09","Type":"ContainerStarted","Data":"6362be6d36f803ed4aa66382c87948cf2cb59136318d621cb188f9a05136116d"} Mar 13 09:35:27 crc kubenswrapper[4930]: I0313 09:35:27.103292 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:27 crc kubenswrapper[4930]: I0313 09:35:27.111672 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc498063-054b-4708-981d-65da056e5b2f","Type":"ContainerStarted","Data":"27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac"} Mar 13 09:35:27 crc kubenswrapper[4930]: I0313 09:35:27.114900 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75f4c75545-5xgvn" event={"ID":"46162d09-ed14-4445-9644-3658980a85e5","Type":"ContainerStarted","Data":"fd3c9d70f863f68a0bf19b72bc1d6f52b31080d53458ab7d5707ee070d62f4c4"} Mar 13 09:35:27 crc kubenswrapper[4930]: I0313 09:35:27.114960 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75f4c75545-5xgvn" event={"ID":"46162d09-ed14-4445-9644-3658980a85e5","Type":"ContainerStarted","Data":"213366d61bb4b346ff73c16826eb4cf31e87aad8cb957ca389cc30d3cd6f3583"} Mar 13 09:35:27 crc kubenswrapper[4930]: I0313 09:35:27.114971 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75f4c75545-5xgvn" event={"ID":"46162d09-ed14-4445-9644-3658980a85e5","Type":"ContainerStarted","Data":"15ea41b8e513ae20cd6034b6d1933360ad7c5775601fdd9151e3bb594cdaecc3"} Mar 13 09:35:27 crc kubenswrapper[4930]: I0313 09:35:27.121801 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" podStartSLOduration=6.121782397 podStartE2EDuration="6.121782397s" podCreationTimestamp="2026-03-13 09:35:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:27.120615596 +0000 UTC m=+1367.870530273" watchObservedRunningTime="2026-03-13 09:35:27.121782397 +0000 UTC m=+1367.871697074" Mar 13 09:35:27 crc kubenswrapper[4930]: I0313 09:35:27.150573 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-75f4c75545-5xgvn" podStartSLOduration=3.15055306 podStartE2EDuration="3.15055306s" podCreationTimestamp="2026-03-13 09:35:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:27.14292003 +0000 UTC m=+1367.892834707" watchObservedRunningTime="2026-03-13 09:35:27.15055306 +0000 UTC m=+1367.900467727" Mar 13 09:35:28 crc kubenswrapper[4930]: I0313 09:35:28.125832 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:29 crc kubenswrapper[4930]: I0313 09:35:29.137051 4930 generic.go:334] "Generic (PLEG): container finished" podID="10e08b2d-bcbf-4b39-9baa-e4ca5213567c" containerID="cb71701e4d8a02633aae086604550da618517bad71a9e78fdb788997e04e82c5" exitCode=0 Mar 13 09:35:29 crc kubenswrapper[4930]: I0313 09:35:29.137336 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-djnqj" event={"ID":"10e08b2d-bcbf-4b39-9baa-e4ca5213567c","Type":"ContainerDied","Data":"cb71701e4d8a02633aae086604550da618517bad71a9e78fdb788997e04e82c5"} Mar 13 09:35:29 crc kubenswrapper[4930]: I0313 09:35:29.139769 4930 generic.go:334] "Generic (PLEG): container finished" podID="66e9286f-5c5e-49e5-9952-8d01782f2a09" containerID="c959109492edd2baef172e975037522c981f72caba726ee20c38339cc13fd18d" exitCode=0 Mar 13 09:35:29 crc kubenswrapper[4930]: I0313 09:35:29.140925 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqk65" event={"ID":"66e9286f-5c5e-49e5-9952-8d01782f2a09","Type":"ContainerDied","Data":"c959109492edd2baef172e975037522c981f72caba726ee20c38339cc13fd18d"} Mar 13 09:35:30 crc kubenswrapper[4930]: I0313 09:35:30.153512 4930 generic.go:334] "Generic (PLEG): container finished" podID="d3fb676a-c7df-4e03-a4ad-41da78bde051" containerID="ac95d58ba906ac3a4c40acd96be960d3b3dfc8dacf73f640c3c779d6f7b101df" exitCode=0 Mar 13 09:35:30 crc kubenswrapper[4930]: I0313 09:35:30.153981 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h9bf6" event={"ID":"d3fb676a-c7df-4e03-a4ad-41da78bde051","Type":"ContainerDied","Data":"ac95d58ba906ac3a4c40acd96be960d3b3dfc8dacf73f640c3c779d6f7b101df"} Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.348230 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-djnqj" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.357602 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.376216 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.376254 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.458718 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-scripts\") pod \"66e9286f-5c5e-49e5-9952-8d01782f2a09\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.459041 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-combined-ca-bundle\") pod \"66e9286f-5c5e-49e5-9952-8d01782f2a09\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.459187 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-fernet-keys\") pod \"66e9286f-5c5e-49e5-9952-8d01782f2a09\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.459342 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-config-data\") pod \"66e9286f-5c5e-49e5-9952-8d01782f2a09\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.459588 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-logs\") pod \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.459873 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-credential-keys\") pod \"66e9286f-5c5e-49e5-9952-8d01782f2a09\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.460237 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-scripts\") pod \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.460680 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-config-data\") pod \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.460869 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-logs" (OuterVolumeSpecName: "logs") pod "10e08b2d-bcbf-4b39-9baa-e4ca5213567c" (UID: "10e08b2d-bcbf-4b39-9baa-e4ca5213567c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.460980 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vfkm8\" (UniqueName: \"kubernetes.io/projected/66e9286f-5c5e-49e5-9952-8d01782f2a09-kube-api-access-vfkm8\") pod \"66e9286f-5c5e-49e5-9952-8d01782f2a09\" (UID: \"66e9286f-5c5e-49e5-9952-8d01782f2a09\") " Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.461378 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-combined-ca-bundle\") pod \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.461603 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqwxf\" (UniqueName: \"kubernetes.io/projected/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-kube-api-access-bqwxf\") pod \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\" (UID: \"10e08b2d-bcbf-4b39-9baa-e4ca5213567c\") " Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.466757 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "66e9286f-5c5e-49e5-9952-8d01782f2a09" (UID: "66e9286f-5c5e-49e5-9952-8d01782f2a09"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.467040 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.474071 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66e9286f-5c5e-49e5-9952-8d01782f2a09-kube-api-access-vfkm8" (OuterVolumeSpecName: "kube-api-access-vfkm8") pod "66e9286f-5c5e-49e5-9952-8d01782f2a09" (UID: "66e9286f-5c5e-49e5-9952-8d01782f2a09"). InnerVolumeSpecName "kube-api-access-vfkm8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.474145 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-scripts" (OuterVolumeSpecName: "scripts") pod "66e9286f-5c5e-49e5-9952-8d01782f2a09" (UID: "66e9286f-5c5e-49e5-9952-8d01782f2a09"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.474200 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "66e9286f-5c5e-49e5-9952-8d01782f2a09" (UID: "66e9286f-5c5e-49e5-9952-8d01782f2a09"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.474245 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-scripts" (OuterVolumeSpecName: "scripts") pod "10e08b2d-bcbf-4b39-9baa-e4ca5213567c" (UID: "10e08b2d-bcbf-4b39-9baa-e4ca5213567c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.474243 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-kube-api-access-bqwxf" (OuterVolumeSpecName: "kube-api-access-bqwxf") pod "10e08b2d-bcbf-4b39-9baa-e4ca5213567c" (UID: "10e08b2d-bcbf-4b39-9baa-e4ca5213567c"). InnerVolumeSpecName "kube-api-access-bqwxf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.504187 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.504665 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.538600 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-config-data" (OuterVolumeSpecName: "config-data") pod "10e08b2d-bcbf-4b39-9baa-e4ca5213567c" (UID: "10e08b2d-bcbf-4b39-9baa-e4ca5213567c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.555562 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-config-data" (OuterVolumeSpecName: "config-data") pod "66e9286f-5c5e-49e5-9952-8d01782f2a09" (UID: "66e9286f-5c5e-49e5-9952-8d01782f2a09"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.570365 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.570408 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vfkm8\" (UniqueName: \"kubernetes.io/projected/66e9286f-5c5e-49e5-9952-8d01782f2a09-kube-api-access-vfkm8\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.570424 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqwxf\" (UniqueName: \"kubernetes.io/projected/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-kube-api-access-bqwxf\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.570452 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.570466 4930 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.570477 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.570488 4930 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.570499 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.584625 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66e9286f-5c5e-49e5-9952-8d01782f2a09" (UID: "66e9286f-5c5e-49e5-9952-8d01782f2a09"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.589244 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10e08b2d-bcbf-4b39-9baa-e4ca5213567c" (UID: "10e08b2d-bcbf-4b39-9baa-e4ca5213567c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.672947 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10e08b2d-bcbf-4b39-9baa-e4ca5213567c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:31 crc kubenswrapper[4930]: I0313 09:35:31.672982 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66e9286f-5c5e-49e5-9952-8d01782f2a09-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.131798 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.194140 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-mqk65" event={"ID":"66e9286f-5c5e-49e5-9952-8d01782f2a09","Type":"ContainerDied","Data":"b36f7adb10d8edcf64730260c9c0706254531d64b1a252ad91843b80116c2e4e"} Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.194470 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b36f7adb10d8edcf64730260c9c0706254531d64b1a252ad91843b80116c2e4e" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.194564 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-mqk65" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.209646 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-g6rkm"] Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.209944 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" podUID="d6f38ccc-56eb-4a6e-88ab-321f81d721d5" containerName="dnsmasq-dns" containerID="cri-o://dd92ca79209c20efba50eff1a6182a6fde22fa49443435bce612d221c18eb1d0" gracePeriod=10 Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.220845 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-djnqj" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.223590 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-djnqj" event={"ID":"10e08b2d-bcbf-4b39-9baa-e4ca5213567c","Type":"ContainerDied","Data":"02a0e5980658806ec9b3ca5294d033a490a0be1174d565afeb30a636c3aa2ef3"} Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.223650 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="02a0e5980658806ec9b3ca5294d033a490a0be1174d565afeb30a636c3aa2ef3" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.223683 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.224169 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.228857 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" podUID="d6f38ccc-56eb-4a6e-88ab-321f81d721d5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.194:5353: connect: connection reset by peer" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.475886 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.477131 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.477464 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.477575 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.489783 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6fd55ffb6d-npc7j"] Mar 13 09:35:32 crc kubenswrapper[4930]: E0313 09:35:32.490341 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66e9286f-5c5e-49e5-9952-8d01782f2a09" containerName="keystone-bootstrap" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.490367 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="66e9286f-5c5e-49e5-9952-8d01782f2a09" containerName="keystone-bootstrap" Mar 13 09:35:32 crc kubenswrapper[4930]: E0313 09:35:32.490393 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10e08b2d-bcbf-4b39-9baa-e4ca5213567c" containerName="placement-db-sync" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.490402 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="10e08b2d-bcbf-4b39-9baa-e4ca5213567c" containerName="placement-db-sync" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.490875 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="66e9286f-5c5e-49e5-9952-8d01782f2a09" containerName="keystone-bootstrap" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.490908 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="10e08b2d-bcbf-4b39-9baa-e4ca5213567c" containerName="placement-db-sync" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.492894 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.497122 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.497555 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.497860 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.498111 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-z6lh5" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.514725 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6fd55ffb6d-npc7j"] Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.529613 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.584057 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.602130 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afc45a41-5494-41d9-90af-67a584dac2eb-logs\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.602198 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vl5sz\" (UniqueName: \"kubernetes.io/projected/afc45a41-5494-41d9-90af-67a584dac2eb-kube-api-access-vl5sz\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.602246 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-public-tls-certs\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.602394 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-combined-ca-bundle\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.602458 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-internal-tls-certs\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.602490 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-config-data\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.602556 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-scripts\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.610257 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-c7549c7dc-lw9qh"] Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.611732 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.612422 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.620779 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-ws54h" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.621170 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.621455 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.621577 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.621654 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.621790 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.642563 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c7549c7dc-lw9qh"] Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705059 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-internal-tls-certs\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705113 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-config-data\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705159 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-internal-tls-certs\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705205 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-combined-ca-bundle\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705232 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-scripts\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705252 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hmkb\" (UniqueName: \"kubernetes.io/projected/60251b90-be5f-4872-ae50-199fd843885c-kube-api-access-8hmkb\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705329 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-scripts\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705402 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afc45a41-5494-41d9-90af-67a584dac2eb-logs\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705444 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-public-tls-certs\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705469 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vl5sz\" (UniqueName: \"kubernetes.io/projected/afc45a41-5494-41d9-90af-67a584dac2eb-kube-api-access-vl5sz\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705501 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-public-tls-certs\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705538 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-credential-keys\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705564 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-fernet-keys\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705613 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-combined-ca-bundle\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.705633 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-config-data\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.707136 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afc45a41-5494-41d9-90af-67a584dac2eb-logs\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.712534 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-scripts\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.712898 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-public-tls-certs\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.721433 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-config-data\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.728284 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vl5sz\" (UniqueName: \"kubernetes.io/projected/afc45a41-5494-41d9-90af-67a584dac2eb-kube-api-access-vl5sz\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.729168 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-internal-tls-certs\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.730888 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-combined-ca-bundle\") pod \"placement-6fd55ffb6d-npc7j\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.807199 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-config-data\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.807293 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-internal-tls-certs\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.807325 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-combined-ca-bundle\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.807362 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hmkb\" (UniqueName: \"kubernetes.io/projected/60251b90-be5f-4872-ae50-199fd843885c-kube-api-access-8hmkb\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.807459 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-scripts\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.807522 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-public-tls-certs\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.807582 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-credential-keys\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.807609 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-fernet-keys\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.814085 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-fernet-keys\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.814534 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-public-tls-certs\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.815110 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-internal-tls-certs\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.818876 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-config-data\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.819938 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-scripts\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.821073 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-credential-keys\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.821134 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/60251b90-be5f-4872-ae50-199fd843885c-combined-ca-bundle\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.826283 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hmkb\" (UniqueName: \"kubernetes.io/projected/60251b90-be5f-4872-ae50-199fd843885c-kube-api-access-8hmkb\") pod \"keystone-c7549c7dc-lw9qh\" (UID: \"60251b90-be5f-4872-ae50-199fd843885c\") " pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.865689 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:32 crc kubenswrapper[4930]: I0313 09:35:32.941822 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:33 crc kubenswrapper[4930]: I0313 09:35:33.242515 4930 generic.go:334] "Generic (PLEG): container finished" podID="d6f38ccc-56eb-4a6e-88ab-321f81d721d5" containerID="dd92ca79209c20efba50eff1a6182a6fde22fa49443435bce612d221c18eb1d0" exitCode=0 Mar 13 09:35:33 crc kubenswrapper[4930]: I0313 09:35:33.242633 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" event={"ID":"d6f38ccc-56eb-4a6e-88ab-321f81d721d5","Type":"ContainerDied","Data":"dd92ca79209c20efba50eff1a6182a6fde22fa49443435bce612d221c18eb1d0"} Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.258012 4930 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.258639 4930 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.627477 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-57c8b65b96-szpn5"] Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.642145 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.672875 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-57c8b65b96-szpn5"] Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.780637 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-internal-tls-certs\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.780907 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-combined-ca-bundle\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.781187 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-config-data\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.781226 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-scripts\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.781564 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-public-tls-certs\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.781918 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/551bdc8a-fb9c-48f5-8379-2ce71515aa73-logs\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.782114 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bwmf\" (UniqueName: \"kubernetes.io/projected/551bdc8a-fb9c-48f5-8379-2ce71515aa73-kube-api-access-5bwmf\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.884013 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-config-data\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.885171 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-scripts\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.885328 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-public-tls-certs\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.885488 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/551bdc8a-fb9c-48f5-8379-2ce71515aa73-logs\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.885634 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bwmf\" (UniqueName: \"kubernetes.io/projected/551bdc8a-fb9c-48f5-8379-2ce71515aa73-kube-api-access-5bwmf\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.885783 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-internal-tls-certs\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.885966 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-combined-ca-bundle\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.885995 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/551bdc8a-fb9c-48f5-8379-2ce71515aa73-logs\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.889685 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-scripts\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.892859 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-public-tls-certs\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.893219 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-config-data\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.900321 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-combined-ca-bundle\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.902389 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/551bdc8a-fb9c-48f5-8379-2ce71515aa73-internal-tls-certs\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.903060 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bwmf\" (UniqueName: \"kubernetes.io/projected/551bdc8a-fb9c-48f5-8379-2ce71515aa73-kube-api-access-5bwmf\") pod \"placement-57c8b65b96-szpn5\" (UID: \"551bdc8a-fb9c-48f5-8379-2ce71515aa73\") " pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:34 crc kubenswrapper[4930]: I0313 09:35:34.973034 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.291334 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-h9bf6" event={"ID":"d3fb676a-c7df-4e03-a4ad-41da78bde051","Type":"ContainerDied","Data":"c617f8fbba5d3e9be1a4c1e4e7a1c07d1562efcb5c7f5158a487e6b45aa8e2be"} Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.291646 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c617f8fbba5d3e9be1a4c1e4e7a1c07d1562efcb5c7f5158a487e6b45aa8e2be" Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.326912 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.396015 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3fb676a-c7df-4e03-a4ad-41da78bde051-db-sync-config-data\") pod \"d3fb676a-c7df-4e03-a4ad-41da78bde051\" (UID: \"d3fb676a-c7df-4e03-a4ad-41da78bde051\") " Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.396144 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pmwb\" (UniqueName: \"kubernetes.io/projected/d3fb676a-c7df-4e03-a4ad-41da78bde051-kube-api-access-4pmwb\") pod \"d3fb676a-c7df-4e03-a4ad-41da78bde051\" (UID: \"d3fb676a-c7df-4e03-a4ad-41da78bde051\") " Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.396193 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fb676a-c7df-4e03-a4ad-41da78bde051-combined-ca-bundle\") pod \"d3fb676a-c7df-4e03-a4ad-41da78bde051\" (UID: \"d3fb676a-c7df-4e03-a4ad-41da78bde051\") " Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.406628 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fb676a-c7df-4e03-a4ad-41da78bde051-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d3fb676a-c7df-4e03-a4ad-41da78bde051" (UID: "d3fb676a-c7df-4e03-a4ad-41da78bde051"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.406673 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3fb676a-c7df-4e03-a4ad-41da78bde051-kube-api-access-4pmwb" (OuterVolumeSpecName: "kube-api-access-4pmwb") pod "d3fb676a-c7df-4e03-a4ad-41da78bde051" (UID: "d3fb676a-c7df-4e03-a4ad-41da78bde051"). InnerVolumeSpecName "kube-api-access-4pmwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.484835 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d3fb676a-c7df-4e03-a4ad-41da78bde051-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d3fb676a-c7df-4e03-a4ad-41da78bde051" (UID: "d3fb676a-c7df-4e03-a4ad-41da78bde051"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.518387 4930 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d3fb676a-c7df-4e03-a4ad-41da78bde051-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.518422 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pmwb\" (UniqueName: \"kubernetes.io/projected/d3fb676a-c7df-4e03-a4ad-41da78bde051-kube-api-access-4pmwb\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.518464 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d3fb676a-c7df-4e03-a4ad-41da78bde051-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.802947 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.931238 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-dns-swift-storage-0\") pod \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.931702 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-ovsdbserver-nb\") pod \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.933958 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-dns-svc\") pod \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.934400 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dxhlm\" (UniqueName: \"kubernetes.io/projected/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-kube-api-access-dxhlm\") pod \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.934483 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-config\") pod \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.934573 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-ovsdbserver-sb\") pod \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\" (UID: \"d6f38ccc-56eb-4a6e-88ab-321f81d721d5\") " Mar 13 09:35:35 crc kubenswrapper[4930]: I0313 09:35:35.943108 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-kube-api-access-dxhlm" (OuterVolumeSpecName: "kube-api-access-dxhlm") pod "d6f38ccc-56eb-4a6e-88ab-321f81d721d5" (UID: "d6f38ccc-56eb-4a6e-88ab-321f81d721d5"). InnerVolumeSpecName "kube-api-access-dxhlm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.038835 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dxhlm\" (UniqueName: \"kubernetes.io/projected/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-kube-api-access-dxhlm\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.042833 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d6f38ccc-56eb-4a6e-88ab-321f81d721d5" (UID: "d6f38ccc-56eb-4a6e-88ab-321f81d721d5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.056601 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-config" (OuterVolumeSpecName: "config") pod "d6f38ccc-56eb-4a6e-88ab-321f81d721d5" (UID: "d6f38ccc-56eb-4a6e-88ab-321f81d721d5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.088142 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d6f38ccc-56eb-4a6e-88ab-321f81d721d5" (UID: "d6f38ccc-56eb-4a6e-88ab-321f81d721d5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.100970 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d6f38ccc-56eb-4a6e-88ab-321f81d721d5" (UID: "d6f38ccc-56eb-4a6e-88ab-321f81d721d5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.141560 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.141591 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.141602 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.141614 4930 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.189190 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d6f38ccc-56eb-4a6e-88ab-321f81d721d5" (UID: "d6f38ccc-56eb-4a6e-88ab-321f81d721d5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.244400 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d6f38ccc-56eb-4a6e-88ab-321f81d721d5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.258298 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-57c8b65b96-szpn5"] Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.285671 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-c7549c7dc-lw9qh"] Mar 13 09:35:36 crc kubenswrapper[4930]: W0313 09:35:36.288082 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod60251b90_be5f_4872_ae50_199fd843885c.slice/crio-be2e6eda2835058309f4637c38aaa8de8eabdaef41813ad6f2ed1fc4a06f32b8 WatchSource:0}: Error finding container be2e6eda2835058309f4637c38aaa8de8eabdaef41813ad6f2ed1fc4a06f32b8: Status 404 returned error can't find the container with id be2e6eda2835058309f4637c38aaa8de8eabdaef41813ad6f2ed1fc4a06f32b8 Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.304917 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6fd55ffb6d-npc7j"] Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.320700 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-57c8b65b96-szpn5" event={"ID":"551bdc8a-fb9c-48f5-8379-2ce71515aa73","Type":"ContainerStarted","Data":"6687a1bfd21343486bf382f68b4eaa66eecd38bdf8538006d4eb850c761e0a83"} Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.323211 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" event={"ID":"d6f38ccc-56eb-4a6e-88ab-321f81d721d5","Type":"ContainerDied","Data":"3be41492351932246cfe5e38c9ccf6d92ed4ca9e400b8c76579d01c4671732bf"} Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.323254 4930 scope.go:117] "RemoveContainer" containerID="dd92ca79209c20efba50eff1a6182a6fde22fa49443435bce612d221c18eb1d0" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.323387 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-g6rkm" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.325473 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c7549c7dc-lw9qh" event={"ID":"60251b90-be5f-4872-ae50-199fd843885c","Type":"ContainerStarted","Data":"be2e6eda2835058309f4637c38aaa8de8eabdaef41813ad6f2ed1fc4a06f32b8"} Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.331745 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc498063-054b-4708-981d-65da056e5b2f","Type":"ContainerStarted","Data":"809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a"} Mar 13 09:35:36 crc kubenswrapper[4930]: W0313 09:35:36.332446 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podafc45a41_5494_41d9_90af_67a584dac2eb.slice/crio-470b76b6d6bf3e77ed12b3aaac0a9c61e8f2e2775e634961299f7f36bedc3323 WatchSource:0}: Error finding container 470b76b6d6bf3e77ed12b3aaac0a9c61e8f2e2775e634961299f7f36bedc3323: Status 404 returned error can't find the container with id 470b76b6d6bf3e77ed12b3aaac0a9c61e8f2e2775e634961299f7f36bedc3323 Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.335700 4930 generic.go:334] "Generic (PLEG): container finished" podID="ae61cc80-b8f2-401c-87db-5728d566b288" containerID="6e35cb5389d90f31f41ef51b7c6a85b9edc8e6f4e042fbf370a1bba9fd1433f9" exitCode=0 Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.335789 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-h9bf6" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.336751 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-p4nk9" event={"ID":"ae61cc80-b8f2-401c-87db-5728d566b288","Type":"ContainerDied","Data":"6e35cb5389d90f31f41ef51b7c6a85b9edc8e6f4e042fbf370a1bba9fd1433f9"} Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.422514 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-g6rkm"] Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.426393 4930 scope.go:117] "RemoveContainer" containerID="49c88ce438a76d30e7936c57f380667742b09a6215e9d1d7f78054f1e44091c9" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.489103 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-g6rkm"] Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.722128 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-f8568957f-mcjmh"] Mar 13 09:35:36 crc kubenswrapper[4930]: E0313 09:35:36.723567 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3fb676a-c7df-4e03-a4ad-41da78bde051" containerName="barbican-db-sync" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.723587 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3fb676a-c7df-4e03-a4ad-41da78bde051" containerName="barbican-db-sync" Mar 13 09:35:36 crc kubenswrapper[4930]: E0313 09:35:36.723649 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f38ccc-56eb-4a6e-88ab-321f81d721d5" containerName="init" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.723659 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f38ccc-56eb-4a6e-88ab-321f81d721d5" containerName="init" Mar 13 09:35:36 crc kubenswrapper[4930]: E0313 09:35:36.723683 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d6f38ccc-56eb-4a6e-88ab-321f81d721d5" containerName="dnsmasq-dns" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.723692 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d6f38ccc-56eb-4a6e-88ab-321f81d721d5" containerName="dnsmasq-dns" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.724176 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d6f38ccc-56eb-4a6e-88ab-321f81d721d5" containerName="dnsmasq-dns" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.724218 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3fb676a-c7df-4e03-a4ad-41da78bde051" containerName="barbican-db-sync" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.747725 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.761749 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.762001 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-c5wbx" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.782714 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.846504 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-84f66dfdb-r96x2"] Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.864799 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.867348 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mxqd\" (UniqueName: \"kubernetes.io/projected/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-kube-api-access-8mxqd\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.868953 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-logs\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.869164 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-config-data\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.869289 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-config-data-custom\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.869477 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-combined-ca-bundle\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.870742 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.912511 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-f8568957f-mcjmh"] Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.940121 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-84f66dfdb-r96x2"] Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.974224 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f96534b5-f1c4-4871-b730-1285a13f0c77-config-data\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.974588 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bs6rq\" (UniqueName: \"kubernetes.io/projected/f96534b5-f1c4-4871-b730-1285a13f0c77-kube-api-access-bs6rq\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.974637 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-config-data\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.974674 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-config-data-custom\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.974709 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f96534b5-f1c4-4871-b730-1285a13f0c77-combined-ca-bundle\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.974747 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-combined-ca-bundle\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.974987 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mxqd\" (UniqueName: \"kubernetes.io/projected/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-kube-api-access-8mxqd\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.975005 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f96534b5-f1c4-4871-b730-1285a13f0c77-config-data-custom\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.975040 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f96534b5-f1c4-4871-b730-1285a13f0c77-logs\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.975074 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-logs\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:36 crc kubenswrapper[4930]: I0313 09:35:36.975409 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-logs\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.050495 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-h7xpv"] Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.052401 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.078507 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-h7xpv"] Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.083886 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f96534b5-f1c4-4871-b730-1285a13f0c77-config-data\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.087773 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bs6rq\" (UniqueName: \"kubernetes.io/projected/f96534b5-f1c4-4871-b730-1285a13f0c77-kube-api-access-bs6rq\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.088212 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f96534b5-f1c4-4871-b730-1285a13f0c77-combined-ca-bundle\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.088665 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f96534b5-f1c4-4871-b730-1285a13f0c77-config-data-custom\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.088940 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f96534b5-f1c4-4871-b730-1285a13f0c77-logs\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.102730 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f96534b5-f1c4-4871-b730-1285a13f0c77-logs\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.170094 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-combined-ca-bundle\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.178268 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-b65b564db-bw6fx"] Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.189944 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.191028 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-config-data-custom\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.191595 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-config-data\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.192278 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.192349 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.192394 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.192499 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-config\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.192553 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zjpb9\" (UniqueName: \"kubernetes.io/projected/341ae9b2-e088-46e1-8a17-60da5410bff8-kube-api-access-zjpb9\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.192595 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.192899 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.193067 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mxqd\" (UniqueName: \"kubernetes.io/projected/67cb25ed-1336-4136-9e3f-c2ae585e6ce8-kube-api-access-8mxqd\") pod \"barbican-worker-f8568957f-mcjmh\" (UID: \"67cb25ed-1336-4136-9e3f-c2ae585e6ce8\") " pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.198157 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f96534b5-f1c4-4871-b730-1285a13f0c77-config-data\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.198610 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f96534b5-f1c4-4871-b730-1285a13f0c77-combined-ca-bundle\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.198868 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bs6rq\" (UniqueName: \"kubernetes.io/projected/f96534b5-f1c4-4871-b730-1285a13f0c77-kube-api-access-bs6rq\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.199475 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f96534b5-f1c4-4871-b730-1285a13f0c77-config-data-custom\") pod \"barbican-keystone-listener-84f66dfdb-r96x2\" (UID: \"f96534b5-f1c4-4871-b730-1285a13f0c77\") " pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.220207 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-b65b564db-bw6fx"] Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.296111 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zjpb9\" (UniqueName: \"kubernetes.io/projected/341ae9b2-e088-46e1-8a17-60da5410bff8-kube-api-access-zjpb9\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.296189 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-combined-ca-bundle\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.296236 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.296285 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-config-data\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.296319 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8bq4\" (UniqueName: \"kubernetes.io/projected/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-kube-api-access-l8bq4\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.296340 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-config-data-custom\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.296387 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.296418 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-logs\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.296472 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.296504 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.296574 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-config\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.298906 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.299526 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.300085 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.300756 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.311081 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-config\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.340499 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zjpb9\" (UniqueName: \"kubernetes.io/projected/341ae9b2-e088-46e1-8a17-60da5410bff8-kube-api-access-zjpb9\") pod \"dnsmasq-dns-848cf88cfc-h7xpv\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.349712 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fd55ffb6d-npc7j" event={"ID":"afc45a41-5494-41d9-90af-67a584dac2eb","Type":"ContainerStarted","Data":"470b76b6d6bf3e77ed12b3aaac0a9c61e8f2e2775e634961299f7f36bedc3323"} Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.363056 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-57c8b65b96-szpn5" event={"ID":"551bdc8a-fb9c-48f5-8379-2ce71515aa73","Type":"ContainerStarted","Data":"4635185b3a3de5a9d172932d2f16241eee27409a75f4ae17a6ab2d02d02fb4a2"} Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.400422 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-combined-ca-bundle\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.400785 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-config-data\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.400890 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8bq4\" (UniqueName: \"kubernetes.io/projected/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-kube-api-access-l8bq4\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.400982 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-config-data-custom\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.401797 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-logs\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.402530 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-logs\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.412059 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-config-data-custom\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.412667 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-f8568957f-mcjmh" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.413605 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-combined-ca-bundle\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.427566 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-config-data\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.428885 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8bq4\" (UniqueName: \"kubernetes.io/projected/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-kube-api-access-l8bq4\") pod \"barbican-api-b65b564db-bw6fx\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.538519 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.580663 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:37 crc kubenswrapper[4930]: I0313 09:35:37.592681 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:37.993475 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d6f38ccc-56eb-4a6e-88ab-321f81d721d5" path="/var/lib/kubelet/pods/d6f38ccc-56eb-4a6e-88ab-321f81d721d5/volumes" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.086359 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-p4nk9" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.252551 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9m4t\" (UniqueName: \"kubernetes.io/projected/ae61cc80-b8f2-401c-87db-5728d566b288-kube-api-access-d9m4t\") pod \"ae61cc80-b8f2-401c-87db-5728d566b288\" (UID: \"ae61cc80-b8f2-401c-87db-5728d566b288\") " Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.252815 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae61cc80-b8f2-401c-87db-5728d566b288-combined-ca-bundle\") pod \"ae61cc80-b8f2-401c-87db-5728d566b288\" (UID: \"ae61cc80-b8f2-401c-87db-5728d566b288\") " Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.252907 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae61cc80-b8f2-401c-87db-5728d566b288-config-data\") pod \"ae61cc80-b8f2-401c-87db-5728d566b288\" (UID: \"ae61cc80-b8f2-401c-87db-5728d566b288\") " Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.264704 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae61cc80-b8f2-401c-87db-5728d566b288-kube-api-access-d9m4t" (OuterVolumeSpecName: "kube-api-access-d9m4t") pod "ae61cc80-b8f2-401c-87db-5728d566b288" (UID: "ae61cc80-b8f2-401c-87db-5728d566b288"). InnerVolumeSpecName "kube-api-access-d9m4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.309844 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-f8568957f-mcjmh"] Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.340687 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae61cc80-b8f2-401c-87db-5728d566b288-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae61cc80-b8f2-401c-87db-5728d566b288" (UID: "ae61cc80-b8f2-401c-87db-5728d566b288"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.355390 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae61cc80-b8f2-401c-87db-5728d566b288-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.355426 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9m4t\" (UniqueName: \"kubernetes.io/projected/ae61cc80-b8f2-401c-87db-5728d566b288-kube-api-access-d9m4t\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.376167 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae61cc80-b8f2-401c-87db-5728d566b288-config-data" (OuterVolumeSpecName: "config-data") pod "ae61cc80-b8f2-401c-87db-5728d566b288" (UID: "ae61cc80-b8f2-401c-87db-5728d566b288"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.412659 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7dsrx" event={"ID":"98c022d1-9a39-4f3e-8108-e4be2b287077","Type":"ContainerStarted","Data":"d6e741c1a755fc3772e85e18f098c7f5930be5a84a595c19c2f9249b2409cc75"} Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.420279 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-c7549c7dc-lw9qh" event={"ID":"60251b90-be5f-4872-ae50-199fd843885c","Type":"ContainerStarted","Data":"0e29479fc4e02e83fd3aac3398e00187d54502b5cae0aba3a6df3565760c50c5"} Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.422308 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.458279 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae61cc80-b8f2-401c-87db-5728d566b288-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.459141 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fd55ffb6d-npc7j" event={"ID":"afc45a41-5494-41d9-90af-67a584dac2eb","Type":"ContainerStarted","Data":"21b7c4ee405bb6bfc241dca312c28f6dd147ec3b084d0a234ffe5583877eb769"} Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.459186 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fd55ffb6d-npc7j" event={"ID":"afc45a41-5494-41d9-90af-67a584dac2eb","Type":"ContainerStarted","Data":"522d784a063767088689e8af88000689c45e83076a50850a7c9ca1b45432b759"} Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.460144 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.460185 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.474293 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-7dsrx" podStartSLOduration=5.362873332 podStartE2EDuration="1m0.474274546s" podCreationTimestamp="2026-03-13 09:34:38 +0000 UTC" firstStartedPulling="2026-03-13 09:34:40.677271442 +0000 UTC m=+1321.427186129" lastFinishedPulling="2026-03-13 09:35:35.788672666 +0000 UTC m=+1376.538587343" observedRunningTime="2026-03-13 09:35:38.449962989 +0000 UTC m=+1379.199877676" watchObservedRunningTime="2026-03-13 09:35:38.474274546 +0000 UTC m=+1379.224189223" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.505180 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-c7549c7dc-lw9qh" podStartSLOduration=6.505160254 podStartE2EDuration="6.505160254s" podCreationTimestamp="2026-03-13 09:35:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:38.47598509 +0000 UTC m=+1379.225899767" watchObservedRunningTime="2026-03-13 09:35:38.505160254 +0000 UTC m=+1379.255074931" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.515315 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-p4nk9" event={"ID":"ae61cc80-b8f2-401c-87db-5728d566b288","Type":"ContainerDied","Data":"4c2f6d942ca4ca1a6cb8e8011be467e830c740a36c8a97c24b93b563e0766c3f"} Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.515347 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c2f6d942ca4ca1a6cb8e8011be467e830c740a36c8a97c24b93b563e0766c3f" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.515416 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-p4nk9" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.539041 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f8568957f-mcjmh" event={"ID":"67cb25ed-1336-4136-9e3f-c2ae585e6ce8","Type":"ContainerStarted","Data":"500698bcad28f58cba62fc6bc28d96d193eabfea2e128cd3eda7659e68916b47"} Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.544572 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-h7xpv"] Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.560720 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-57c8b65b96-szpn5" event={"ID":"551bdc8a-fb9c-48f5-8379-2ce71515aa73","Type":"ContainerStarted","Data":"d263101a43ba0fa3ffcc81933eea9e3da0b5e49d38b45551bddb9e75d0d69f50"} Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.561590 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.561622 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.596332 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6fd55ffb6d-npc7j" podStartSLOduration=6.59631101 podStartE2EDuration="6.59631101s" podCreationTimestamp="2026-03-13 09:35:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:38.509895148 +0000 UTC m=+1379.259809825" watchObservedRunningTime="2026-03-13 09:35:38.59631101 +0000 UTC m=+1379.346225677" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.611712 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-57c8b65b96-szpn5" podStartSLOduration=4.611691593 podStartE2EDuration="4.611691593s" podCreationTimestamp="2026-03-13 09:35:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:38.58827738 +0000 UTC m=+1379.338192067" watchObservedRunningTime="2026-03-13 09:35:38.611691593 +0000 UTC m=+1379.361606270" Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.761137 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-84f66dfdb-r96x2"] Mar 13 09:35:38 crc kubenswrapper[4930]: I0313 09:35:38.840325 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-b65b564db-bw6fx"] Mar 13 09:35:39 crc kubenswrapper[4930]: I0313 09:35:39.404993 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:39 crc kubenswrapper[4930]: I0313 09:35:39.405536 4930 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 09:35:39 crc kubenswrapper[4930]: I0313 09:35:39.500733 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 09:35:39 crc kubenswrapper[4930]: I0313 09:35:39.500808 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 09:35:39 crc kubenswrapper[4930]: I0313 09:35:39.524278 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 09:35:39 crc kubenswrapper[4930]: I0313 09:35:39.613684 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" event={"ID":"f96534b5-f1c4-4871-b730-1285a13f0c77","Type":"ContainerStarted","Data":"fc929f3d33b9831c4f40517c0a34fe6c048da6c9984067d01d665c96a8390ee3"} Mar 13 09:35:39 crc kubenswrapper[4930]: I0313 09:35:39.644769 4930 generic.go:334] "Generic (PLEG): container finished" podID="341ae9b2-e088-46e1-8a17-60da5410bff8" containerID="f334a9894e9e082fd364616106b0632d7dfda449022cc28b65318b7b2825deb9" exitCode=0 Mar 13 09:35:39 crc kubenswrapper[4930]: I0313 09:35:39.644878 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" event={"ID":"341ae9b2-e088-46e1-8a17-60da5410bff8","Type":"ContainerDied","Data":"f334a9894e9e082fd364616106b0632d7dfda449022cc28b65318b7b2825deb9"} Mar 13 09:35:39 crc kubenswrapper[4930]: I0313 09:35:39.644910 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" event={"ID":"341ae9b2-e088-46e1-8a17-60da5410bff8","Type":"ContainerStarted","Data":"ef9f71e5f9568210ad1699f805bd6047e47e39f1469c16eec78eb8e26225b63b"} Mar 13 09:35:39 crc kubenswrapper[4930]: I0313 09:35:39.681882 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b65b564db-bw6fx" event={"ID":"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61","Type":"ContainerStarted","Data":"037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558"} Mar 13 09:35:39 crc kubenswrapper[4930]: I0313 09:35:39.681925 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b65b564db-bw6fx" event={"ID":"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61","Type":"ContainerStarted","Data":"a35b54cc6a21fe31946ecc41d0300118feaf9bb57e47d8f47d93948350e3dc90"} Mar 13 09:35:40 crc kubenswrapper[4930]: I0313 09:35:40.751486 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" event={"ID":"341ae9b2-e088-46e1-8a17-60da5410bff8","Type":"ContainerStarted","Data":"d5a0405cf24925c4355fe82e0a9fbe9be52796b9ae80109aac1b1c0df179f935"} Mar 13 09:35:40 crc kubenswrapper[4930]: I0313 09:35:40.752124 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:40 crc kubenswrapper[4930]: I0313 09:35:40.766244 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b65b564db-bw6fx" event={"ID":"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61","Type":"ContainerStarted","Data":"bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2"} Mar 13 09:35:40 crc kubenswrapper[4930]: I0313 09:35:40.766287 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:40 crc kubenswrapper[4930]: I0313 09:35:40.766317 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:40 crc kubenswrapper[4930]: I0313 09:35:40.793667 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" podStartSLOduration=4.793645038 podStartE2EDuration="4.793645038s" podCreationTimestamp="2026-03-13 09:35:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:40.782708012 +0000 UTC m=+1381.532622689" watchObservedRunningTime="2026-03-13 09:35:40.793645038 +0000 UTC m=+1381.543559715" Mar 13 09:35:40 crc kubenswrapper[4930]: I0313 09:35:40.818441 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-b65b564db-bw6fx" podStartSLOduration=3.818407647 podStartE2EDuration="3.818407647s" podCreationTimestamp="2026-03-13 09:35:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:40.815619444 +0000 UTC m=+1381.565534121" watchObservedRunningTime="2026-03-13 09:35:40.818407647 +0000 UTC m=+1381.568322324" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.250732 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-7f55767858-jxftm"] Mar 13 09:35:41 crc kubenswrapper[4930]: E0313 09:35:41.251664 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae61cc80-b8f2-401c-87db-5728d566b288" containerName="heat-db-sync" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.251687 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae61cc80-b8f2-401c-87db-5728d566b288" containerName="heat-db-sync" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.251982 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae61cc80-b8f2-401c-87db-5728d566b288" containerName="heat-db-sync" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.253503 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.256477 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.256713 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.277786 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f55767858-jxftm"] Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.315088 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/313c3b7e-2c50-4583-9dee-c141274c239a-logs\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.315177 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9d2zj\" (UniqueName: \"kubernetes.io/projected/313c3b7e-2c50-4583-9dee-c141274c239a-kube-api-access-9d2zj\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.315214 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-internal-tls-certs\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.315268 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-config-data-custom\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.315313 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-public-tls-certs\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.315329 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-config-data\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.315401 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-combined-ca-bundle\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.417458 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9d2zj\" (UniqueName: \"kubernetes.io/projected/313c3b7e-2c50-4583-9dee-c141274c239a-kube-api-access-9d2zj\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.417526 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-internal-tls-certs\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.417592 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-config-data-custom\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.417640 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-public-tls-certs\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.417659 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-config-data\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.417732 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-combined-ca-bundle\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.417793 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/313c3b7e-2c50-4583-9dee-c141274c239a-logs\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.418324 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/313c3b7e-2c50-4583-9dee-c141274c239a-logs\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.424917 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-config-data-custom\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.429060 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-config-data\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.431010 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-combined-ca-bundle\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.436216 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-internal-tls-certs\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.443247 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9d2zj\" (UniqueName: \"kubernetes.io/projected/313c3b7e-2c50-4583-9dee-c141274c239a-kube-api-access-9d2zj\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.444461 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/313c3b7e-2c50-4583-9dee-c141274c239a-public-tls-certs\") pod \"barbican-api-7f55767858-jxftm\" (UID: \"313c3b7e-2c50-4583-9dee-c141274c239a\") " pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:41 crc kubenswrapper[4930]: I0313 09:35:41.576215 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:43 crc kubenswrapper[4930]: I0313 09:35:43.317423 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-7f55767858-jxftm"] Mar 13 09:35:43 crc kubenswrapper[4930]: I0313 09:35:43.823330 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f55767858-jxftm" event={"ID":"313c3b7e-2c50-4583-9dee-c141274c239a","Type":"ContainerStarted","Data":"420bc3d54ef1fcb875b576c4a8bce5643b45be4156fd8dd18c3348e38c48ef2f"} Mar 13 09:35:43 crc kubenswrapper[4930]: I0313 09:35:43.823652 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f55767858-jxftm" event={"ID":"313c3b7e-2c50-4583-9dee-c141274c239a","Type":"ContainerStarted","Data":"6cc89ad2bf7ff6819cda4f60f39d1793df319d9b93ce02a8496074641667ffe6"} Mar 13 09:35:43 crc kubenswrapper[4930]: I0313 09:35:43.825738 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" event={"ID":"f96534b5-f1c4-4871-b730-1285a13f0c77","Type":"ContainerStarted","Data":"229e6e945f0248ab929f68d374e37dc11c88fccc1f55221e4f65505d3b34f29f"} Mar 13 09:35:43 crc kubenswrapper[4930]: I0313 09:35:43.825770 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" event={"ID":"f96534b5-f1c4-4871-b730-1285a13f0c77","Type":"ContainerStarted","Data":"7d7b8c568580a63df00c8491cb89a224c82a8f2df2cb5cb8853e33b7ef340bf5"} Mar 13 09:35:43 crc kubenswrapper[4930]: I0313 09:35:43.829238 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f8568957f-mcjmh" event={"ID":"67cb25ed-1336-4136-9e3f-c2ae585e6ce8","Type":"ContainerStarted","Data":"16d38561713eccc71c112fa384e00c0f0b332eae420eb7f5986eb6fc52ee071b"} Mar 13 09:35:43 crc kubenswrapper[4930]: I0313 09:35:43.829284 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-f8568957f-mcjmh" event={"ID":"67cb25ed-1336-4136-9e3f-c2ae585e6ce8","Type":"ContainerStarted","Data":"b26051ed503374d137efe1cc9acae2e44003bb68a24b5a0e0edb62e1bb02ddde"} Mar 13 09:35:43 crc kubenswrapper[4930]: I0313 09:35:43.852790 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-84f66dfdb-r96x2" podStartSLOduration=3.816333026 podStartE2EDuration="7.852772236s" podCreationTimestamp="2026-03-13 09:35:36 +0000 UTC" firstStartedPulling="2026-03-13 09:35:38.746473171 +0000 UTC m=+1379.496387858" lastFinishedPulling="2026-03-13 09:35:42.782912391 +0000 UTC m=+1383.532827068" observedRunningTime="2026-03-13 09:35:43.842337803 +0000 UTC m=+1384.592252500" watchObservedRunningTime="2026-03-13 09:35:43.852772236 +0000 UTC m=+1384.602686913" Mar 13 09:35:43 crc kubenswrapper[4930]: I0313 09:35:43.866574 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-f8568957f-mcjmh" podStartSLOduration=3.426967584 podStartE2EDuration="7.866549927s" podCreationTimestamp="2026-03-13 09:35:36 +0000 UTC" firstStartedPulling="2026-03-13 09:35:38.32964706 +0000 UTC m=+1379.079561737" lastFinishedPulling="2026-03-13 09:35:42.769229403 +0000 UTC m=+1383.519144080" observedRunningTime="2026-03-13 09:35:43.859532943 +0000 UTC m=+1384.609447620" watchObservedRunningTime="2026-03-13 09:35:43.866549927 +0000 UTC m=+1384.616464604" Mar 13 09:35:44 crc kubenswrapper[4930]: I0313 09:35:44.843760 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-7f55767858-jxftm" event={"ID":"313c3b7e-2c50-4583-9dee-c141274c239a","Type":"ContainerStarted","Data":"bceb33d5fba64494d918008142a808d1e62878208d66dd92caa20723f7084e86"} Mar 13 09:35:44 crc kubenswrapper[4930]: I0313 09:35:44.844643 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:44 crc kubenswrapper[4930]: I0313 09:35:44.844678 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:44 crc kubenswrapper[4930]: I0313 09:35:44.869088 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-7f55767858-jxftm" podStartSLOduration=3.869066909 podStartE2EDuration="3.869066909s" podCreationTimestamp="2026-03-13 09:35:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:44.864790907 +0000 UTC m=+1385.614705584" watchObservedRunningTime="2026-03-13 09:35:44.869066909 +0000 UTC m=+1385.618981586" Mar 13 09:35:47 crc kubenswrapper[4930]: I0313 09:35:47.582591 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:35:47 crc kubenswrapper[4930]: I0313 09:35:47.658380 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-5hc9m"] Mar 13 09:35:47 crc kubenswrapper[4930]: I0313 09:35:47.661308 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" podUID="9a434276-b75a-465b-b5d8-d75dc7624d09" containerName="dnsmasq-dns" containerID="cri-o://6362be6d36f803ed4aa66382c87948cf2cb59136318d621cb188f9a05136116d" gracePeriod=10 Mar 13 09:35:47 crc kubenswrapper[4930]: I0313 09:35:47.882922 4930 generic.go:334] "Generic (PLEG): container finished" podID="9a434276-b75a-465b-b5d8-d75dc7624d09" containerID="6362be6d36f803ed4aa66382c87948cf2cb59136318d621cb188f9a05136116d" exitCode=0 Mar 13 09:35:47 crc kubenswrapper[4930]: I0313 09:35:47.882966 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" event={"ID":"9a434276-b75a-465b-b5d8-d75dc7624d09","Type":"ContainerDied","Data":"6362be6d36f803ed4aa66382c87948cf2cb59136318d621cb188f9a05136116d"} Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.212420 4930 generic.go:334] "Generic (PLEG): container finished" podID="98c022d1-9a39-4f3e-8108-e4be2b287077" containerID="d6e741c1a755fc3772e85e18f098c7f5930be5a84a595c19c2f9249b2409cc75" exitCode=0 Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.213175 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7dsrx" event={"ID":"98c022d1-9a39-4f3e-8108-e4be2b287077","Type":"ContainerDied","Data":"d6e741c1a755fc3772e85e18f098c7f5930be5a84a595c19c2f9249b2409cc75"} Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.572797 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.650789 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kw6j\" (UniqueName: \"kubernetes.io/projected/9a434276-b75a-465b-b5d8-d75dc7624d09-kube-api-access-2kw6j\") pod \"9a434276-b75a-465b-b5d8-d75dc7624d09\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.651303 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-dns-swift-storage-0\") pod \"9a434276-b75a-465b-b5d8-d75dc7624d09\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.651543 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-dns-svc\") pod \"9a434276-b75a-465b-b5d8-d75dc7624d09\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.652239 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-ovsdbserver-sb\") pod \"9a434276-b75a-465b-b5d8-d75dc7624d09\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.652403 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-config\") pod \"9a434276-b75a-465b-b5d8-d75dc7624d09\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.652550 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-ovsdbserver-nb\") pod \"9a434276-b75a-465b-b5d8-d75dc7624d09\" (UID: \"9a434276-b75a-465b-b5d8-d75dc7624d09\") " Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.658800 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a434276-b75a-465b-b5d8-d75dc7624d09-kube-api-access-2kw6j" (OuterVolumeSpecName: "kube-api-access-2kw6j") pod "9a434276-b75a-465b-b5d8-d75dc7624d09" (UID: "9a434276-b75a-465b-b5d8-d75dc7624d09"). InnerVolumeSpecName "kube-api-access-2kw6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.755343 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kw6j\" (UniqueName: \"kubernetes.io/projected/9a434276-b75a-465b-b5d8-d75dc7624d09-kube-api-access-2kw6j\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.777139 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9a434276-b75a-465b-b5d8-d75dc7624d09" (UID: "9a434276-b75a-465b-b5d8-d75dc7624d09"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.779639 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9a434276-b75a-465b-b5d8-d75dc7624d09" (UID: "9a434276-b75a-465b-b5d8-d75dc7624d09"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.780056 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9a434276-b75a-465b-b5d8-d75dc7624d09" (UID: "9a434276-b75a-465b-b5d8-d75dc7624d09"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.789221 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9a434276-b75a-465b-b5d8-d75dc7624d09" (UID: "9a434276-b75a-465b-b5d8-d75dc7624d09"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.795643 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-config" (OuterVolumeSpecName: "config") pod "9a434276-b75a-465b-b5d8-d75dc7624d09" (UID: "9a434276-b75a-465b-b5d8-d75dc7624d09"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.858791 4930 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.858835 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.858848 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.858859 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.858871 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9a434276-b75a-465b-b5d8-d75dc7624d09-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.936841 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.954906 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-b65b564db-bw6fx" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerName="barbican-api" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:35:50 crc kubenswrapper[4930]: I0313 09:35:50.958679 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-b65b564db-bw6fx" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerName="barbican-api-log" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.223486 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.223523 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7b667979-5hc9m" event={"ID":"9a434276-b75a-465b-b5d8-d75dc7624d09","Type":"ContainerDied","Data":"1bcfa8ff35ca2fdbd05533fa07cf290e83c83b3f1b3a0f4636fc2a3c11cef7dc"} Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.223961 4930 scope.go:117] "RemoveContainer" containerID="6362be6d36f803ed4aa66382c87948cf2cb59136318d621cb188f9a05136116d" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.226280 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc498063-054b-4708-981d-65da056e5b2f","Type":"ContainerStarted","Data":"76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e"} Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.226402 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="ceilometer-central-agent" containerID="cri-o://02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff" gracePeriod=30 Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.226486 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="sg-core" containerID="cri-o://809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a" gracePeriod=30 Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.226514 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="ceilometer-notification-agent" containerID="cri-o://27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac" gracePeriod=30 Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.226551 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="proxy-httpd" containerID="cri-o://76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e" gracePeriod=30 Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.257010 4930 scope.go:117] "RemoveContainer" containerID="2f4c8976574acf8e6b97bfb6522062d106f8fa43e8bd739b4e739aaef12953ec" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.283734 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.958349836 podStartE2EDuration="1m12.283713772s" podCreationTimestamp="2026-03-13 09:34:39 +0000 UTC" firstStartedPulling="2026-03-13 09:34:40.964599503 +0000 UTC m=+1321.714514180" lastFinishedPulling="2026-03-13 09:35:50.289963439 +0000 UTC m=+1391.039878116" observedRunningTime="2026-03-13 09:35:51.269660454 +0000 UTC m=+1392.019575151" watchObservedRunningTime="2026-03-13 09:35:51.283713772 +0000 UTC m=+1392.033628449" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.300025 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-5hc9m"] Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.312677 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7b667979-5hc9m"] Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.644865 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.686441 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-scripts\") pod \"98c022d1-9a39-4f3e-8108-e4be2b287077\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.686580 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-combined-ca-bundle\") pod \"98c022d1-9a39-4f3e-8108-e4be2b287077\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.686638 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/98c022d1-9a39-4f3e-8108-e4be2b287077-etc-machine-id\") pod \"98c022d1-9a39-4f3e-8108-e4be2b287077\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.686710 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-config-data\") pod \"98c022d1-9a39-4f3e-8108-e4be2b287077\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.686805 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmcpt\" (UniqueName: \"kubernetes.io/projected/98c022d1-9a39-4f3e-8108-e4be2b287077-kube-api-access-gmcpt\") pod \"98c022d1-9a39-4f3e-8108-e4be2b287077\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.686838 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-db-sync-config-data\") pod \"98c022d1-9a39-4f3e-8108-e4be2b287077\" (UID: \"98c022d1-9a39-4f3e-8108-e4be2b287077\") " Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.687576 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/98c022d1-9a39-4f3e-8108-e4be2b287077-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "98c022d1-9a39-4f3e-8108-e4be2b287077" (UID: "98c022d1-9a39-4f3e-8108-e4be2b287077"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.694789 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-scripts" (OuterVolumeSpecName: "scripts") pod "98c022d1-9a39-4f3e-8108-e4be2b287077" (UID: "98c022d1-9a39-4f3e-8108-e4be2b287077"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.697822 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "98c022d1-9a39-4f3e-8108-e4be2b287077" (UID: "98c022d1-9a39-4f3e-8108-e4be2b287077"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.704290 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98c022d1-9a39-4f3e-8108-e4be2b287077-kube-api-access-gmcpt" (OuterVolumeSpecName: "kube-api-access-gmcpt") pod "98c022d1-9a39-4f3e-8108-e4be2b287077" (UID: "98c022d1-9a39-4f3e-8108-e4be2b287077"). InnerVolumeSpecName "kube-api-access-gmcpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.755361 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98c022d1-9a39-4f3e-8108-e4be2b287077" (UID: "98c022d1-9a39-4f3e-8108-e4be2b287077"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.765269 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-config-data" (OuterVolumeSpecName: "config-data") pod "98c022d1-9a39-4f3e-8108-e4be2b287077" (UID: "98c022d1-9a39-4f3e-8108-e4be2b287077"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.789657 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmcpt\" (UniqueName: \"kubernetes.io/projected/98c022d1-9a39-4f3e-8108-e4be2b287077-kube-api-access-gmcpt\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.789712 4930 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.789723 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.789732 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.789740 4930 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/98c022d1-9a39-4f3e-8108-e4be2b287077-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.789750 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98c022d1-9a39-4f3e-8108-e4be2b287077-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.904074 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-g6hb8"] Mar 13 09:35:51 crc kubenswrapper[4930]: E0313 09:35:51.904691 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98c022d1-9a39-4f3e-8108-e4be2b287077" containerName="cinder-db-sync" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.904714 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="98c022d1-9a39-4f3e-8108-e4be2b287077" containerName="cinder-db-sync" Mar 13 09:35:51 crc kubenswrapper[4930]: E0313 09:35:51.904728 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a434276-b75a-465b-b5d8-d75dc7624d09" containerName="dnsmasq-dns" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.904736 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a434276-b75a-465b-b5d8-d75dc7624d09" containerName="dnsmasq-dns" Mar 13 09:35:51 crc kubenswrapper[4930]: E0313 09:35:51.904759 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a434276-b75a-465b-b5d8-d75dc7624d09" containerName="init" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.904770 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a434276-b75a-465b-b5d8-d75dc7624d09" containerName="init" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.905020 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="98c022d1-9a39-4f3e-8108-e4be2b287077" containerName="cinder-db-sync" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.905046 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a434276-b75a-465b-b5d8-d75dc7624d09" containerName="dnsmasq-dns" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.906659 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.915268 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g6hb8"] Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.983049 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a434276-b75a-465b-b5d8-d75dc7624d09" path="/var/lib/kubelet/pods/9a434276-b75a-465b-b5d8-d75dc7624d09/volumes" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.993639 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdpzp\" (UniqueName: \"kubernetes.io/projected/2b4f2c08-b04f-40e6-a90c-256225efe861-kube-api-access-zdpzp\") pod \"community-operators-g6hb8\" (UID: \"2b4f2c08-b04f-40e6-a90c-256225efe861\") " pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.993778 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b4f2c08-b04f-40e6-a90c-256225efe861-utilities\") pod \"community-operators-g6hb8\" (UID: \"2b4f2c08-b04f-40e6-a90c-256225efe861\") " pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:35:51 crc kubenswrapper[4930]: I0313 09:35:51.993856 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b4f2c08-b04f-40e6-a90c-256225efe861-catalog-content\") pod \"community-operators-g6hb8\" (UID: \"2b4f2c08-b04f-40e6-a90c-256225efe861\") " pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.097241 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdpzp\" (UniqueName: \"kubernetes.io/projected/2b4f2c08-b04f-40e6-a90c-256225efe861-kube-api-access-zdpzp\") pod \"community-operators-g6hb8\" (UID: \"2b4f2c08-b04f-40e6-a90c-256225efe861\") " pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.097352 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b4f2c08-b04f-40e6-a90c-256225efe861-utilities\") pod \"community-operators-g6hb8\" (UID: \"2b4f2c08-b04f-40e6-a90c-256225efe861\") " pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.097449 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b4f2c08-b04f-40e6-a90c-256225efe861-catalog-content\") pod \"community-operators-g6hb8\" (UID: \"2b4f2c08-b04f-40e6-a90c-256225efe861\") " pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.098368 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b4f2c08-b04f-40e6-a90c-256225efe861-catalog-content\") pod \"community-operators-g6hb8\" (UID: \"2b4f2c08-b04f-40e6-a90c-256225efe861\") " pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.100478 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b4f2c08-b04f-40e6-a90c-256225efe861-utilities\") pod \"community-operators-g6hb8\" (UID: \"2b4f2c08-b04f-40e6-a90c-256225efe861\") " pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.116401 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdpzp\" (UniqueName: \"kubernetes.io/projected/2b4f2c08-b04f-40e6-a90c-256225efe861-kube-api-access-zdpzp\") pod \"community-operators-g6hb8\" (UID: \"2b4f2c08-b04f-40e6-a90c-256225efe861\") " pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.238555 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-7dsrx" event={"ID":"98c022d1-9a39-4f3e-8108-e4be2b287077","Type":"ContainerDied","Data":"9d82304b36b5b9ad7d85fbaf6cd66db99326cb32098ccb87ecd939ee6cee8456"} Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.238919 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d82304b36b5b9ad7d85fbaf6cd66db99326cb32098ccb87ecd939ee6cee8456" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.238711 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.238607 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-7dsrx" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.240142 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.242236 4930 generic.go:334] "Generic (PLEG): container finished" podID="bc498063-054b-4708-981d-65da056e5b2f" containerID="76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e" exitCode=0 Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.242286 4930 generic.go:334] "Generic (PLEG): container finished" podID="bc498063-054b-4708-981d-65da056e5b2f" containerID="809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a" exitCode=2 Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.242296 4930 generic.go:334] "Generic (PLEG): container finished" podID="bc498063-054b-4708-981d-65da056e5b2f" containerID="02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff" exitCode=0 Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.242374 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc498063-054b-4708-981d-65da056e5b2f","Type":"ContainerDied","Data":"76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e"} Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.242401 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc498063-054b-4708-981d-65da056e5b2f","Type":"ContainerDied","Data":"809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a"} Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.242417 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc498063-054b-4708-981d-65da056e5b2f","Type":"ContainerDied","Data":"02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff"} Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.709643 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.712058 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.720709 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-p6m7w" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.721023 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.721100 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.725396 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.732552 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.776155 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-75f4c75545-5xgvn"] Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.776492 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-75f4c75545-5xgvn" podUID="46162d09-ed14-4445-9644-3658980a85e5" containerName="neutron-api" containerID="cri-o://213366d61bb4b346ff73c16826eb4cf31e87aad8cb957ca389cc30d3cd6f3583" gracePeriod=30 Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.776980 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-75f4c75545-5xgvn" podUID="46162d09-ed14-4445-9644-3658980a85e5" containerName="neutron-httpd" containerID="cri-o://fd3c9d70f863f68a0bf19b72bc1d6f52b31080d53458ab7d5707ee070d62f4c4" gracePeriod=30 Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.801530 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-75f4c75545-5xgvn" podUID="46162d09-ed14-4445-9644-3658980a85e5" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.202:9696/\": read tcp 10.217.0.2:34754->10.217.0.202:9696: read: connection reset by peer" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.853570 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5cc4fbb59f-99vq6"] Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.855603 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.865372 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5cc4fbb59f-99vq6"] Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.867791 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-scripts\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.867905 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.868018 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.868065 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-config-data\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.868098 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnswt\" (UniqueName: \"kubernetes.io/projected/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-kube-api-access-pnswt\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.868216 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.878870 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7wx75"] Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.881643 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.906544 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7wx75"] Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970038 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970343 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970377 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-config-data\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970401 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnswt\" (UniqueName: \"kubernetes.io/projected/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-kube-api-access-pnswt\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970445 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-httpd-config\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970478 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-ovndb-tls-certs\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970515 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970540 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-combined-ca-bundle\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970562 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970598 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x9h6b\" (UniqueName: \"kubernetes.io/projected/c765eaa6-eb16-4c09-b845-430cfd29e6d6-kube-api-access-x9h6b\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970613 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-dns-svc\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970636 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-config\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970666 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lf999\" (UniqueName: \"kubernetes.io/projected/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-kube-api-access-lf999\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970688 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-scripts\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970703 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-internal-tls-certs\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970740 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970755 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-public-tls-certs\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970783 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.970803 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-config\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.971852 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.981606 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-config-data\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.984861 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.987502 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.993461 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 09:35:52 crc kubenswrapper[4930]: I0313 09:35:52.996286 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.000916 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.000950 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-scripts\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.004624 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnswt\" (UniqueName: \"kubernetes.io/projected/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-kube-api-access-pnswt\") pod \"cinder-scheduler-0\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " pod="openstack/cinder-scheduler-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.013183 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.072746 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.072802 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-config-data\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.072828 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-logs\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.072887 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-httpd-config\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.072922 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-ovndb-tls-certs\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.072950 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-config-data-custom\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.072981 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073008 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-combined-ca-bundle\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073029 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-scripts\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073043 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073078 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073096 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x9h6b\" (UniqueName: \"kubernetes.io/projected/c765eaa6-eb16-4c09-b845-430cfd29e6d6-kube-api-access-x9h6b\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073126 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-dns-svc\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073151 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-config\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073187 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lf999\" (UniqueName: \"kubernetes.io/projected/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-kube-api-access-lf999\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073223 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-internal-tls-certs\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073261 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrtxs\" (UniqueName: \"kubernetes.io/projected/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-kube-api-access-zrtxs\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073287 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-public-tls-certs\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073343 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073387 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-config\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.073710 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.075191 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.076542 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-dns-svc\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.084367 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-config\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.100902 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.101056 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.101534 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-httpd-config\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.109645 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-public-tls-certs\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.110979 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-config\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.114057 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-internal-tls-certs\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.117034 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-combined-ca-bundle\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.124099 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-ovndb-tls-certs\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.130125 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lf999\" (UniqueName: \"kubernetes.io/projected/c871e8e3-f075-4b8f-ad91-6552b7ec66b0-kube-api-access-lf999\") pod \"neutron-5cc4fbb59f-99vq6\" (UID: \"c871e8e3-f075-4b8f-ad91-6552b7ec66b0\") " pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.151391 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x9h6b\" (UniqueName: \"kubernetes.io/projected/c765eaa6-eb16-4c09-b845-430cfd29e6d6-kube-api-access-x9h6b\") pod \"dnsmasq-dns-6578955fd5-7wx75\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.177002 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-scripts\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.177048 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.177096 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.177199 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrtxs\" (UniqueName: \"kubernetes.io/projected/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-kube-api-access-zrtxs\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.177321 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-config-data\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.177347 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-logs\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.177455 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-config-data-custom\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.177548 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.178548 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-logs\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.199825 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.214709 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g6hb8"] Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.216150 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.223187 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.223482 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-config-data\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.232001 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-config-data-custom\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.232075 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-scripts\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.234839 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrtxs\" (UniqueName: \"kubernetes.io/projected/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-kube-api-access-zrtxs\") pod \"cinder-api-0\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.302929 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g6hb8" event={"ID":"2b4f2c08-b04f-40e6-a90c-256225efe861","Type":"ContainerStarted","Data":"d7da926ac2d9ba108fb69324c81c5b1a78c95f6a49da6eaeb895a0f55e033be5"} Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.328359 4930 generic.go:334] "Generic (PLEG): container finished" podID="46162d09-ed14-4445-9644-3658980a85e5" containerID="fd3c9d70f863f68a0bf19b72bc1d6f52b31080d53458ab7d5707ee070d62f4c4" exitCode=0 Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.328403 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75f4c75545-5xgvn" event={"ID":"46162d09-ed14-4445-9644-3658980a85e5","Type":"ContainerDied","Data":"fd3c9d70f863f68a0bf19b72bc1d6f52b31080d53458ab7d5707ee070d62f4c4"} Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.366581 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.875201 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:53 crc kubenswrapper[4930]: I0313 09:35:53.975092 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 09:35:54 crc kubenswrapper[4930]: I0313 09:35:54.143727 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7wx75"] Mar 13 09:35:54 crc kubenswrapper[4930]: I0313 09:35:54.299574 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5cc4fbb59f-99vq6"] Mar 13 09:35:54 crc kubenswrapper[4930]: I0313 09:35:54.301367 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:54 crc kubenswrapper[4930]: I0313 09:35:54.335827 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 09:35:54 crc kubenswrapper[4930]: I0313 09:35:54.369508 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" event={"ID":"c765eaa6-eb16-4c09-b845-430cfd29e6d6","Type":"ContainerStarted","Data":"b63c6918069f1a6e5d341ac72e2774ff2e089353f795668beb943fe8bb147485"} Mar 13 09:35:54 crc kubenswrapper[4930]: I0313 09:35:54.392927 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cc4fbb59f-99vq6" event={"ID":"c871e8e3-f075-4b8f-ad91-6552b7ec66b0","Type":"ContainerStarted","Data":"a56521a2fe9bbc411fa1a8e9f8da3bec57da6e95bc78507796606a8998ab80db"} Mar 13 09:35:54 crc kubenswrapper[4930]: I0313 09:35:54.430606 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b","Type":"ContainerStarted","Data":"795e2d664423128de17093f5ab139b5579d4ea9629f5b3d24fb35e27040ae23a"} Mar 13 09:35:54 crc kubenswrapper[4930]: I0313 09:35:54.445897 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7f00bfbf-551d-4730-bfeb-4c317e2a60e3","Type":"ContainerStarted","Data":"cfd90ac2096510a024278595d7cad533c17b686aa897e007a3cd76090289c2b7"} Mar 13 09:35:54 crc kubenswrapper[4930]: I0313 09:35:54.480828 4930 generic.go:334] "Generic (PLEG): container finished" podID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerID="e9b75150c047afd99ffe8d248526ab230c1c325dd118d491cac47e43ea408ec2" exitCode=0 Mar 13 09:35:54 crc kubenswrapper[4930]: I0313 09:35:54.480868 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g6hb8" event={"ID":"2b4f2c08-b04f-40e6-a90c-256225efe861","Type":"ContainerDied","Data":"e9b75150c047afd99ffe8d248526ab230c1c325dd118d491cac47e43ea408ec2"} Mar 13 09:35:54 crc kubenswrapper[4930]: I0313 09:35:54.854708 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-75f4c75545-5xgvn" podUID="46162d09-ed14-4445-9644-3658980a85e5" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.202:9696/\": dial tcp 10.217.0.202:9696: connect: connection refused" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.321132 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.442442 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.519760 4930 generic.go:334] "Generic (PLEG): container finished" podID="c765eaa6-eb16-4c09-b845-430cfd29e6d6" containerID="3472ada2eb325f5c27cfe7c71ef733160d23d45e2fdf21f36a61234fc7167d3d" exitCode=0 Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.519810 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" event={"ID":"c765eaa6-eb16-4c09-b845-430cfd29e6d6","Type":"ContainerDied","Data":"3472ada2eb325f5c27cfe7c71ef733160d23d45e2fdf21f36a61234fc7167d3d"} Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.527751 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wrssc\" (UniqueName: \"kubernetes.io/projected/bc498063-054b-4708-981d-65da056e5b2f-kube-api-access-wrssc\") pod \"bc498063-054b-4708-981d-65da056e5b2f\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.527809 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc498063-054b-4708-981d-65da056e5b2f-run-httpd\") pod \"bc498063-054b-4708-981d-65da056e5b2f\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.527910 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-config-data\") pod \"bc498063-054b-4708-981d-65da056e5b2f\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.527957 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-scripts\") pod \"bc498063-054b-4708-981d-65da056e5b2f\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.527979 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-sg-core-conf-yaml\") pod \"bc498063-054b-4708-981d-65da056e5b2f\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.528052 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc498063-054b-4708-981d-65da056e5b2f-log-httpd\") pod \"bc498063-054b-4708-981d-65da056e5b2f\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.528086 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-combined-ca-bundle\") pod \"bc498063-054b-4708-981d-65da056e5b2f\" (UID: \"bc498063-054b-4708-981d-65da056e5b2f\") " Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.529350 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cc4fbb59f-99vq6" event={"ID":"c871e8e3-f075-4b8f-ad91-6552b7ec66b0","Type":"ContainerStarted","Data":"718cfe56cf00f656d332bfa36825e1573699bbee57b9626f76d4c28f3cbb844d"} Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.529725 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc498063-054b-4708-981d-65da056e5b2f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bc498063-054b-4708-981d-65da056e5b2f" (UID: "bc498063-054b-4708-981d-65da056e5b2f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.530135 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc498063-054b-4708-981d-65da056e5b2f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bc498063-054b-4708-981d-65da056e5b2f" (UID: "bc498063-054b-4708-981d-65da056e5b2f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.545783 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc498063-054b-4708-981d-65da056e5b2f-kube-api-access-wrssc" (OuterVolumeSpecName: "kube-api-access-wrssc") pod "bc498063-054b-4708-981d-65da056e5b2f" (UID: "bc498063-054b-4708-981d-65da056e5b2f"). InnerVolumeSpecName "kube-api-access-wrssc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.546191 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-scripts" (OuterVolumeSpecName: "scripts") pod "bc498063-054b-4708-981d-65da056e5b2f" (UID: "bc498063-054b-4708-981d-65da056e5b2f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.580561 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bc498063-054b-4708-981d-65da056e5b2f" (UID: "bc498063-054b-4708-981d-65da056e5b2f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.595720 4930 generic.go:334] "Generic (PLEG): container finished" podID="bc498063-054b-4708-981d-65da056e5b2f" containerID="27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac" exitCode=0 Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.595813 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc498063-054b-4708-981d-65da056e5b2f","Type":"ContainerDied","Data":"27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac"} Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.595841 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bc498063-054b-4708-981d-65da056e5b2f","Type":"ContainerDied","Data":"e531cf6f90563bcb0e755872c6ae653ff60002fbbdd0eda5902d1ae27c193589"} Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.595859 4930 scope.go:117] "RemoveContainer" containerID="76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.596032 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.636766 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wrssc\" (UniqueName: \"kubernetes.io/projected/bc498063-054b-4708-981d-65da056e5b2f-kube-api-access-wrssc\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.636796 4930 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc498063-054b-4708-981d-65da056e5b2f-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.636807 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.636816 4930 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.636826 4930 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bc498063-054b-4708-981d-65da056e5b2f-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.689682 4930 scope.go:117] "RemoveContainer" containerID="809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.764132 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-7f55767858-jxftm" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.824854 4930 scope.go:117] "RemoveContainer" containerID="27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.871051 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-b65b564db-bw6fx"] Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.877347 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-b65b564db-bw6fx" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerName="barbican-api-log" containerID="cri-o://037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558" gracePeriod=30 Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.878821 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-b65b564db-bw6fx" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerName="barbican-api" containerID="cri-o://bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2" gracePeriod=30 Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.905299 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bsl74"] Mar 13 09:35:55 crc kubenswrapper[4930]: E0313 09:35:55.905806 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="proxy-httpd" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.905825 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="proxy-httpd" Mar 13 09:35:55 crc kubenswrapper[4930]: E0313 09:35:55.905843 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="ceilometer-central-agent" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.905849 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="ceilometer-central-agent" Mar 13 09:35:55 crc kubenswrapper[4930]: E0313 09:35:55.905857 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="sg-core" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.905864 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="sg-core" Mar 13 09:35:55 crc kubenswrapper[4930]: E0313 09:35:55.905901 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="ceilometer-notification-agent" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.905908 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="ceilometer-notification-agent" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.906095 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="ceilometer-central-agent" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.906111 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="sg-core" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.906127 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="proxy-httpd" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.906139 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc498063-054b-4708-981d-65da056e5b2f" containerName="ceilometer-notification-agent" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.911469 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.943539 4930 scope.go:117] "RemoveContainer" containerID="02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.947830 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bsl74"] Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.948628 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-catalog-content\") pod \"redhat-operators-bsl74\" (UID: \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\") " pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.948720 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-utilities\") pod \"redhat-operators-bsl74\" (UID: \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\") " pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:35:55 crc kubenswrapper[4930]: I0313 09:35:55.948846 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svgvd\" (UniqueName: \"kubernetes.io/projected/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-kube-api-access-svgvd\") pod \"redhat-operators-bsl74\" (UID: \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\") " pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.049792 4930 scope.go:117] "RemoveContainer" containerID="76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.051654 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svgvd\" (UniqueName: \"kubernetes.io/projected/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-kube-api-access-svgvd\") pod \"redhat-operators-bsl74\" (UID: \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\") " pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.051987 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-catalog-content\") pod \"redhat-operators-bsl74\" (UID: \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\") " pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.052027 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-utilities\") pod \"redhat-operators-bsl74\" (UID: \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\") " pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:35:56 crc kubenswrapper[4930]: E0313 09:35:56.052038 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e\": container with ID starting with 76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e not found: ID does not exist" containerID="76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.052070 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e"} err="failed to get container status \"76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e\": rpc error: code = NotFound desc = could not find container \"76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e\": container with ID starting with 76365ac3cd93c708d445f6b009f7ffdc03edbfb1561c45a3d8d01887d607444e not found: ID does not exist" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.052094 4930 scope.go:117] "RemoveContainer" containerID="809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.052753 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-catalog-content\") pod \"redhat-operators-bsl74\" (UID: \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\") " pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.052935 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-utilities\") pod \"redhat-operators-bsl74\" (UID: \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\") " pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:35:56 crc kubenswrapper[4930]: E0313 09:35:56.053103 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a\": container with ID starting with 809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a not found: ID does not exist" containerID="809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.053133 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a"} err="failed to get container status \"809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a\": rpc error: code = NotFound desc = could not find container \"809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a\": container with ID starting with 809f6a96a0270ad0a9efaf9885b69a75a1f0b83b31d4fa0f5cf685183b76e66a not found: ID does not exist" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.053156 4930 scope.go:117] "RemoveContainer" containerID="27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac" Mar 13 09:35:56 crc kubenswrapper[4930]: E0313 09:35:56.054856 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac\": container with ID starting with 27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac not found: ID does not exist" containerID="27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.054891 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac"} err="failed to get container status \"27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac\": rpc error: code = NotFound desc = could not find container \"27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac\": container with ID starting with 27f38789dcabc7a05fd32a41053a581353681439e4153fd225e64562831e43ac not found: ID does not exist" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.054921 4930 scope.go:117] "RemoveContainer" containerID="02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff" Mar 13 09:35:56 crc kubenswrapper[4930]: E0313 09:35:56.055183 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff\": container with ID starting with 02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff not found: ID does not exist" containerID="02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.055216 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff"} err="failed to get container status \"02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff\": rpc error: code = NotFound desc = could not find container \"02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff\": container with ID starting with 02c2bd173eef6499b450602e588fa0f0efb1e08a89355737f8499ecb3ca996ff not found: ID does not exist" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.077838 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svgvd\" (UniqueName: \"kubernetes.io/projected/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-kube-api-access-svgvd\") pod \"redhat-operators-bsl74\" (UID: \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\") " pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.151008 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bc498063-054b-4708-981d-65da056e5b2f" (UID: "bc498063-054b-4708-981d-65da056e5b2f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.155295 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.190076 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-config-data" (OuterVolumeSpecName: "config-data") pod "bc498063-054b-4708-981d-65da056e5b2f" (UID: "bc498063-054b-4708-981d-65da056e5b2f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.249176 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.259378 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bc498063-054b-4708-981d-65da056e5b2f-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.499714 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.550876 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.652150 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.661675 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.663543 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.664307 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.668521 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.710836 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5cc4fbb59f-99vq6" event={"ID":"c871e8e3-f075-4b8f-ad91-6552b7ec66b0","Type":"ContainerStarted","Data":"7bd7f18b2a5a18afd59f46f4989f83c576f8460eec19a6c20193c153ef0f5814"} Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.711331 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.715768 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7f00bfbf-551d-4730-bfeb-4c317e2a60e3","Type":"ContainerStarted","Data":"74f48ea43202380d568031e44173017842714b290a809a1368e8e7b7a9ab22fa"} Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.733690 4930 generic.go:334] "Generic (PLEG): container finished" podID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerID="037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558" exitCode=143 Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.733759 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b65b564db-bw6fx" event={"ID":"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61","Type":"ContainerDied","Data":"037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558"} Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.807190 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.808044 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.808105 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-scripts\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.808239 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-config-data\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.808292 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1a220dc-49c4-4eb4-b05a-a07f3263b119-log-httpd\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.808362 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f72lc\" (UniqueName: \"kubernetes.io/projected/b1a220dc-49c4-4eb4-b05a-a07f3263b119-kube-api-access-f72lc\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.808411 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1a220dc-49c4-4eb4-b05a-a07f3263b119-run-httpd\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.910547 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-scripts\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.910642 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-config-data\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.910674 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1a220dc-49c4-4eb4-b05a-a07f3263b119-log-httpd\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.910719 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f72lc\" (UniqueName: \"kubernetes.io/projected/b1a220dc-49c4-4eb4-b05a-a07f3263b119-kube-api-access-f72lc\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.910746 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1a220dc-49c4-4eb4-b05a-a07f3263b119-run-httpd\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.910792 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.910847 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.911619 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1a220dc-49c4-4eb4-b05a-a07f3263b119-log-httpd\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.913866 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1a220dc-49c4-4eb4-b05a-a07f3263b119-run-httpd\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.918148 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-scripts\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.918263 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.919140 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-config-data\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.922935 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:56 crc kubenswrapper[4930]: I0313 09:35:56.940009 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f72lc\" (UniqueName: \"kubernetes.io/projected/b1a220dc-49c4-4eb4-b05a-a07f3263b119-kube-api-access-f72lc\") pod \"ceilometer-0\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " pod="openstack/ceilometer-0" Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.031041 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.039553 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5cc4fbb59f-99vq6" podStartSLOduration=5.039531689 podStartE2EDuration="5.039531689s" podCreationTimestamp="2026-03-13 09:35:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:56.771626086 +0000 UTC m=+1397.521540763" watchObservedRunningTime="2026-03-13 09:35:57.039531689 +0000 UTC m=+1397.789446366" Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.054563 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bsl74"] Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.671669 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.762533 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7f00bfbf-551d-4730-bfeb-4c317e2a60e3","Type":"ContainerStarted","Data":"07254621feced3ddd66802379857c0ce63823b857833b6250527f980d8badd16"} Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.762697 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerName="cinder-api-log" containerID="cri-o://74f48ea43202380d568031e44173017842714b290a809a1368e8e7b7a9ab22fa" gracePeriod=30 Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.762998 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.763052 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerName="cinder-api" containerID="cri-o://07254621feced3ddd66802379857c0ce63823b857833b6250527f980d8badd16" gracePeriod=30 Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.776059 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1a220dc-49c4-4eb4-b05a-a07f3263b119","Type":"ContainerStarted","Data":"5431d673d3d4cb2dd906e6abf20ec3740b0f4201cf87905cf5a8d64eb14630fa"} Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.793621 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" event={"ID":"c765eaa6-eb16-4c09-b845-430cfd29e6d6","Type":"ContainerStarted","Data":"ff2905cd51a663f0a9d93f7c4daaf6804a2402ddf3b658878cb899db630b23cd"} Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.794488 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.794419189 podStartE2EDuration="5.794419189s" podCreationTimestamp="2026-03-13 09:35:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:57.78031234 +0000 UTC m=+1398.530227017" watchObservedRunningTime="2026-03-13 09:35:57.794419189 +0000 UTC m=+1398.544333866" Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.794732 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.797062 4930 generic.go:334] "Generic (PLEG): container finished" podID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerID="96782db3f1a05ce85427f059e78463d435d238cf3a17f8dfe64c138a48f73674" exitCode=0 Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.797124 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bsl74" event={"ID":"a73c1bd6-dd06-4e14-b8ee-1916942e53f5","Type":"ContainerDied","Data":"96782db3f1a05ce85427f059e78463d435d238cf3a17f8dfe64c138a48f73674"} Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.797145 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bsl74" event={"ID":"a73c1bd6-dd06-4e14-b8ee-1916942e53f5","Type":"ContainerStarted","Data":"0352d4207fd23d7f82439fb2d1db720a090328debdaa9c88d25c5404d09edf7f"} Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.811120 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b","Type":"ContainerStarted","Data":"b35c75ca056b85198e0f020035b73c2431be9e9688290e0263c297ecee19416b"} Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.826017 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" podStartSLOduration=5.825997906 podStartE2EDuration="5.825997906s" podCreationTimestamp="2026-03-13 09:35:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:35:57.818260803 +0000 UTC m=+1398.568175480" watchObservedRunningTime="2026-03-13 09:35:57.825997906 +0000 UTC m=+1398.575912583" Mar 13 09:35:57 crc kubenswrapper[4930]: I0313 09:35:57.993875 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc498063-054b-4708-981d-65da056e5b2f" path="/var/lib/kubelet/pods/bc498063-054b-4708-981d-65da056e5b2f/volumes" Mar 13 09:35:58 crc kubenswrapper[4930]: I0313 09:35:58.847140 4930 generic.go:334] "Generic (PLEG): container finished" podID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerID="74f48ea43202380d568031e44173017842714b290a809a1368e8e7b7a9ab22fa" exitCode=143 Mar 13 09:35:58 crc kubenswrapper[4930]: I0313 09:35:58.847313 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7f00bfbf-551d-4730-bfeb-4c317e2a60e3","Type":"ContainerDied","Data":"74f48ea43202380d568031e44173017842714b290a809a1368e8e7b7a9ab22fa"} Mar 13 09:35:58 crc kubenswrapper[4930]: I0313 09:35:58.866406 4930 generic.go:334] "Generic (PLEG): container finished" podID="46162d09-ed14-4445-9644-3658980a85e5" containerID="213366d61bb4b346ff73c16826eb4cf31e87aad8cb957ca389cc30d3cd6f3583" exitCode=0 Mar 13 09:35:58 crc kubenswrapper[4930]: I0313 09:35:58.866530 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75f4c75545-5xgvn" event={"ID":"46162d09-ed14-4445-9644-3658980a85e5","Type":"ContainerDied","Data":"213366d61bb4b346ff73c16826eb4cf31e87aad8cb957ca389cc30d3cd6f3583"} Mar 13 09:35:58 crc kubenswrapper[4930]: I0313 09:35:58.871532 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b","Type":"ContainerStarted","Data":"83025acec25b4f21dca183700c7fccc9b6dd4bdf036942fbcf3ca67e0246932c"} Mar 13 09:35:58 crc kubenswrapper[4930]: I0313 09:35:58.896386 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=5.572180351 podStartE2EDuration="6.896195139s" podCreationTimestamp="2026-03-13 09:35:52 +0000 UTC" firstStartedPulling="2026-03-13 09:35:53.994617693 +0000 UTC m=+1394.744532370" lastFinishedPulling="2026-03-13 09:35:55.318632481 +0000 UTC m=+1396.068547158" observedRunningTime="2026-03-13 09:35:58.888051946 +0000 UTC m=+1399.637966623" watchObservedRunningTime="2026-03-13 09:35:58.896195139 +0000 UTC m=+1399.646109816" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.192833 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-b65b564db-bw6fx" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.209:9311/healthcheck\": read tcp 10.217.0.2:54994->10.217.0.209:9311: read: connection reset by peer" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.193248 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-b65b564db-bw6fx" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.209:9311/healthcheck\": read tcp 10.217.0.2:55006->10.217.0.209:9311: read: connection reset by peer" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.286799 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.474279 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-config\") pod \"46162d09-ed14-4445-9644-3658980a85e5\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.474686 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-ovndb-tls-certs\") pod \"46162d09-ed14-4445-9644-3658980a85e5\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.474774 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-public-tls-certs\") pod \"46162d09-ed14-4445-9644-3658980a85e5\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.475178 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-combined-ca-bundle\") pod \"46162d09-ed14-4445-9644-3658980a85e5\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.475299 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnks8\" (UniqueName: \"kubernetes.io/projected/46162d09-ed14-4445-9644-3658980a85e5-kube-api-access-rnks8\") pod \"46162d09-ed14-4445-9644-3658980a85e5\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.475465 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-internal-tls-certs\") pod \"46162d09-ed14-4445-9644-3658980a85e5\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.475559 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-httpd-config\") pod \"46162d09-ed14-4445-9644-3658980a85e5\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.501652 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "46162d09-ed14-4445-9644-3658980a85e5" (UID: "46162d09-ed14-4445-9644-3658980a85e5"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.501816 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46162d09-ed14-4445-9644-3658980a85e5-kube-api-access-rnks8" (OuterVolumeSpecName: "kube-api-access-rnks8") pod "46162d09-ed14-4445-9644-3658980a85e5" (UID: "46162d09-ed14-4445-9644-3658980a85e5"). InnerVolumeSpecName "kube-api-access-rnks8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.577265 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "46162d09-ed14-4445-9644-3658980a85e5" (UID: "46162d09-ed14-4445-9644-3658980a85e5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.579036 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-internal-tls-certs\") pod \"46162d09-ed14-4445-9644-3658980a85e5\" (UID: \"46162d09-ed14-4445-9644-3658980a85e5\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.580286 4930 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.580310 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnks8\" (UniqueName: \"kubernetes.io/projected/46162d09-ed14-4445-9644-3658980a85e5-kube-api-access-rnks8\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:59 crc kubenswrapper[4930]: W0313 09:35:59.580389 4930 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/46162d09-ed14-4445-9644-3658980a85e5/volumes/kubernetes.io~secret/internal-tls-certs Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.580401 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "46162d09-ed14-4445-9644-3658980a85e5" (UID: "46162d09-ed14-4445-9644-3658980a85e5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.587651 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-config" (OuterVolumeSpecName: "config") pod "46162d09-ed14-4445-9644-3658980a85e5" (UID: "46162d09-ed14-4445-9644-3658980a85e5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.591320 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46162d09-ed14-4445-9644-3658980a85e5" (UID: "46162d09-ed14-4445-9644-3658980a85e5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.620887 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "46162d09-ed14-4445-9644-3658980a85e5" (UID: "46162d09-ed14-4445-9644-3658980a85e5"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.681882 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.681910 4930 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.681922 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.681930 4930 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.682311 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "46162d09-ed14-4445-9644-3658980a85e5" (UID: "46162d09-ed14-4445-9644-3658980a85e5"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.785016 4930 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46162d09-ed14-4445-9644-3658980a85e5-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.819815 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.886398 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-config-data\") pod \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.886536 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-logs\") pod \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.886660 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8bq4\" (UniqueName: \"kubernetes.io/projected/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-kube-api-access-l8bq4\") pod \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.886772 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-combined-ca-bundle\") pod \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.886791 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-config-data-custom\") pod \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\" (UID: \"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61\") " Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.888546 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-logs" (OuterVolumeSpecName: "logs") pod "8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" (UID: "8f84d0e4-d0ba-43f7-830c-c9841e8fdc61"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.892391 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" (UID: "8f84d0e4-d0ba-43f7-830c-c9841e8fdc61"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.892630 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1a220dc-49c4-4eb4-b05a-a07f3263b119","Type":"ContainerStarted","Data":"b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf"} Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.892679 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1a220dc-49c4-4eb4-b05a-a07f3263b119","Type":"ContainerStarted","Data":"fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b"} Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.897053 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-kube-api-access-l8bq4" (OuterVolumeSpecName: "kube-api-access-l8bq4") pod "8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" (UID: "8f84d0e4-d0ba-43f7-830c-c9841e8fdc61"). InnerVolumeSpecName "kube-api-access-l8bq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.909698 4930 generic.go:334] "Generic (PLEG): container finished" podID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerID="bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2" exitCode=0 Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.909782 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b65b564db-bw6fx" event={"ID":"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61","Type":"ContainerDied","Data":"bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2"} Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.909810 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-b65b564db-bw6fx" event={"ID":"8f84d0e4-d0ba-43f7-830c-c9841e8fdc61","Type":"ContainerDied","Data":"a35b54cc6a21fe31946ecc41d0300118feaf9bb57e47d8f47d93948350e3dc90"} Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.909825 4930 scope.go:117] "RemoveContainer" containerID="bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.909971 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-b65b564db-bw6fx" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.924853 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-75f4c75545-5xgvn" event={"ID":"46162d09-ed14-4445-9644-3658980a85e5","Type":"ContainerDied","Data":"15ea41b8e513ae20cd6034b6d1933360ad7c5775601fdd9151e3bb594cdaecc3"} Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.924878 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-75f4c75545-5xgvn" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.929497 4930 generic.go:334] "Generic (PLEG): container finished" podID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerID="3f4bcb33f157da3b03c1334f17a6c246e3568f1a75cd98ca29fce9755dba763c" exitCode=0 Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.930652 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bsl74" event={"ID":"a73c1bd6-dd06-4e14-b8ee-1916942e53f5","Type":"ContainerDied","Data":"3f4bcb33f157da3b03c1334f17a6c246e3568f1a75cd98ca29fce9755dba763c"} Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.935890 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" (UID: "8f84d0e4-d0ba-43f7-830c-c9841e8fdc61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.948540 4930 scope.go:117] "RemoveContainer" containerID="037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.991088 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.991120 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8bq4\" (UniqueName: \"kubernetes.io/projected/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-kube-api-access-l8bq4\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.991269 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.991280 4930 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.995590 4930 scope.go:117] "RemoveContainer" containerID="bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2" Mar 13 09:35:59 crc kubenswrapper[4930]: E0313 09:35:59.998093 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2\": container with ID starting with bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2 not found: ID does not exist" containerID="bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.998146 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2"} err="failed to get container status \"bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2\": rpc error: code = NotFound desc = could not find container \"bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2\": container with ID starting with bd095f171acb18d0421a2b5e17f2a7e216f026ece4d10e7439c41551d8fb16e2 not found: ID does not exist" Mar 13 09:35:59 crc kubenswrapper[4930]: I0313 09:35:59.998177 4930 scope.go:117] "RemoveContainer" containerID="037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558" Mar 13 09:36:00 crc kubenswrapper[4930]: E0313 09:36:00.001969 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558\": container with ID starting with 037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558 not found: ID does not exist" containerID="037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.002008 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558"} err="failed to get container status \"037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558\": rpc error: code = NotFound desc = could not find container \"037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558\": container with ID starting with 037edcb3de117b99052dd9c648da3b5f218377767a552ef536b399494166e558 not found: ID does not exist" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.002033 4930 scope.go:117] "RemoveContainer" containerID="fd3c9d70f863f68a0bf19b72bc1d6f52b31080d53458ab7d5707ee070d62f4c4" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.025604 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-config-data" (OuterVolumeSpecName: "config-data") pod "8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" (UID: "8f84d0e4-d0ba-43f7-830c-c9841e8fdc61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.093270 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.165047 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-75f4c75545-5xgvn"] Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.165089 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-75f4c75545-5xgvn"] Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.165111 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556576-kctpg"] Mar 13 09:36:00 crc kubenswrapper[4930]: E0313 09:36:00.166151 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46162d09-ed14-4445-9644-3658980a85e5" containerName="neutron-api" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.166175 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="46162d09-ed14-4445-9644-3658980a85e5" containerName="neutron-api" Mar 13 09:36:00 crc kubenswrapper[4930]: E0313 09:36:00.166211 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerName="barbican-api" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.166221 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerName="barbican-api" Mar 13 09:36:00 crc kubenswrapper[4930]: E0313 09:36:00.166240 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46162d09-ed14-4445-9644-3658980a85e5" containerName="neutron-httpd" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.166248 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="46162d09-ed14-4445-9644-3658980a85e5" containerName="neutron-httpd" Mar 13 09:36:00 crc kubenswrapper[4930]: E0313 09:36:00.166266 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerName="barbican-api-log" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.166273 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerName="barbican-api-log" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.166533 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="46162d09-ed14-4445-9644-3658980a85e5" containerName="neutron-httpd" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.166552 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="46162d09-ed14-4445-9644-3658980a85e5" containerName="neutron-api" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.166564 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerName="barbican-api-log" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.166592 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" containerName="barbican-api" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.167429 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556576-kctpg"] Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.167534 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556576-kctpg" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.169842 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.170153 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.170314 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.194931 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8qcj5\" (UniqueName: \"kubernetes.io/projected/f4ac8095-d474-4fe0-ab66-5e119896d4a1-kube-api-access-8qcj5\") pod \"auto-csr-approver-29556576-kctpg\" (UID: \"f4ac8095-d474-4fe0-ab66-5e119896d4a1\") " pod="openshift-infra/auto-csr-approver-29556576-kctpg" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.213992 4930 scope.go:117] "RemoveContainer" containerID="213366d61bb4b346ff73c16826eb4cf31e87aad8cb957ca389cc30d3cd6f3583" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.253955 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-b65b564db-bw6fx"] Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.265305 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-b65b564db-bw6fx"] Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.296674 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8qcj5\" (UniqueName: \"kubernetes.io/projected/f4ac8095-d474-4fe0-ab66-5e119896d4a1-kube-api-access-8qcj5\") pod \"auto-csr-approver-29556576-kctpg\" (UID: \"f4ac8095-d474-4fe0-ab66-5e119896d4a1\") " pod="openshift-infra/auto-csr-approver-29556576-kctpg" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.312822 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8qcj5\" (UniqueName: \"kubernetes.io/projected/f4ac8095-d474-4fe0-ab66-5e119896d4a1-kube-api-access-8qcj5\") pod \"auto-csr-approver-29556576-kctpg\" (UID: \"f4ac8095-d474-4fe0-ab66-5e119896d4a1\") " pod="openshift-infra/auto-csr-approver-29556576-kctpg" Mar 13 09:36:00 crc kubenswrapper[4930]: I0313 09:36:00.498849 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556576-kctpg" Mar 13 09:36:01 crc kubenswrapper[4930]: I0313 09:36:01.151794 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556576-kctpg"] Mar 13 09:36:01 crc kubenswrapper[4930]: I0313 09:36:01.954107 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556576-kctpg" event={"ID":"f4ac8095-d474-4fe0-ab66-5e119896d4a1","Type":"ContainerStarted","Data":"0805b88195b489aba14cc8e951582338c1d17735d863808e1cb92dbbefa47e38"} Mar 13 09:36:01 crc kubenswrapper[4930]: I0313 09:36:01.984907 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46162d09-ed14-4445-9644-3658980a85e5" path="/var/lib/kubelet/pods/46162d09-ed14-4445-9644-3658980a85e5/volumes" Mar 13 09:36:01 crc kubenswrapper[4930]: I0313 09:36:01.986102 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f84d0e4-d0ba-43f7-830c-c9841e8fdc61" path="/var/lib/kubelet/pods/8f84d0e4-d0ba-43f7-830c-c9841e8fdc61/volumes" Mar 13 09:36:03 crc kubenswrapper[4930]: I0313 09:36:03.100419 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 09:36:03 crc kubenswrapper[4930]: I0313 09:36:03.224887 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:36:03 crc kubenswrapper[4930]: I0313 09:36:03.295369 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-h7xpv"] Mar 13 09:36:03 crc kubenswrapper[4930]: I0313 09:36:03.295637 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" podUID="341ae9b2-e088-46e1-8a17-60da5410bff8" containerName="dnsmasq-dns" containerID="cri-o://d5a0405cf24925c4355fe82e0a9fbe9be52796b9ae80109aac1b1c0df179f935" gracePeriod=10 Mar 13 09:36:04 crc kubenswrapper[4930]: I0313 09:36:04.843703 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:36:04 crc kubenswrapper[4930]: I0313 09:36:04.994527 4930 generic.go:334] "Generic (PLEG): container finished" podID="341ae9b2-e088-46e1-8a17-60da5410bff8" containerID="d5a0405cf24925c4355fe82e0a9fbe9be52796b9ae80109aac1b1c0df179f935" exitCode=0 Mar 13 09:36:04 crc kubenswrapper[4930]: I0313 09:36:04.994606 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" event={"ID":"341ae9b2-e088-46e1-8a17-60da5410bff8","Type":"ContainerDied","Data":"d5a0405cf24925c4355fe82e0a9fbe9be52796b9ae80109aac1b1c0df179f935"} Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.113650 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.408693 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.215:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.595307 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.714855 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-ovsdbserver-nb\") pod \"341ae9b2-e088-46e1-8a17-60da5410bff8\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.714966 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-config\") pod \"341ae9b2-e088-46e1-8a17-60da5410bff8\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.715000 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zjpb9\" (UniqueName: \"kubernetes.io/projected/341ae9b2-e088-46e1-8a17-60da5410bff8-kube-api-access-zjpb9\") pod \"341ae9b2-e088-46e1-8a17-60da5410bff8\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.715023 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-ovsdbserver-sb\") pod \"341ae9b2-e088-46e1-8a17-60da5410bff8\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.715110 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-dns-swift-storage-0\") pod \"341ae9b2-e088-46e1-8a17-60da5410bff8\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.715153 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-dns-svc\") pod \"341ae9b2-e088-46e1-8a17-60da5410bff8\" (UID: \"341ae9b2-e088-46e1-8a17-60da5410bff8\") " Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.724587 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/341ae9b2-e088-46e1-8a17-60da5410bff8-kube-api-access-zjpb9" (OuterVolumeSpecName: "kube-api-access-zjpb9") pod "341ae9b2-e088-46e1-8a17-60da5410bff8" (UID: "341ae9b2-e088-46e1-8a17-60da5410bff8"). InnerVolumeSpecName "kube-api-access-zjpb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.775237 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-config" (OuterVolumeSpecName: "config") pod "341ae9b2-e088-46e1-8a17-60da5410bff8" (UID: "341ae9b2-e088-46e1-8a17-60da5410bff8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.778862 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "341ae9b2-e088-46e1-8a17-60da5410bff8" (UID: "341ae9b2-e088-46e1-8a17-60da5410bff8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.795337 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "341ae9b2-e088-46e1-8a17-60da5410bff8" (UID: "341ae9b2-e088-46e1-8a17-60da5410bff8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.803007 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "341ae9b2-e088-46e1-8a17-60da5410bff8" (UID: "341ae9b2-e088-46e1-8a17-60da5410bff8"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.810760 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "341ae9b2-e088-46e1-8a17-60da5410bff8" (UID: "341ae9b2-e088-46e1-8a17-60da5410bff8"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.817888 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zjpb9\" (UniqueName: \"kubernetes.io/projected/341ae9b2-e088-46e1-8a17-60da5410bff8-kube-api-access-zjpb9\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.817919 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.817930 4930 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.817940 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.817949 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:08 crc kubenswrapper[4930]: I0313 09:36:08.817958 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/341ae9b2-e088-46e1-8a17-60da5410bff8-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:09 crc kubenswrapper[4930]: I0313 09:36:09.039955 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" event={"ID":"341ae9b2-e088-46e1-8a17-60da5410bff8","Type":"ContainerDied","Data":"ef9f71e5f9568210ad1699f805bd6047e47e39f1469c16eec78eb8e26225b63b"} Mar 13 09:36:09 crc kubenswrapper[4930]: I0313 09:36:09.040303 4930 scope.go:117] "RemoveContainer" containerID="d5a0405cf24925c4355fe82e0a9fbe9be52796b9ae80109aac1b1c0df179f935" Mar 13 09:36:09 crc kubenswrapper[4930]: I0313 09:36:09.040015 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" Mar 13 09:36:09 crc kubenswrapper[4930]: I0313 09:36:09.097670 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-h7xpv"] Mar 13 09:36:09 crc kubenswrapper[4930]: I0313 09:36:09.108391 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-h7xpv"] Mar 13 09:36:09 crc kubenswrapper[4930]: I0313 09:36:09.134651 4930 scope.go:117] "RemoveContainer" containerID="f334a9894e9e082fd364616106b0632d7dfda449022cc28b65318b7b2825deb9" Mar 13 09:36:09 crc kubenswrapper[4930]: I0313 09:36:09.988842 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="341ae9b2-e088-46e1-8a17-60da5410bff8" path="/var/lib/kubelet/pods/341ae9b2-e088-46e1-8a17-60da5410bff8/volumes" Mar 13 09:36:10 crc kubenswrapper[4930]: I0313 09:36:10.062451 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1a220dc-49c4-4eb4-b05a-a07f3263b119","Type":"ContainerStarted","Data":"6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb"} Mar 13 09:36:12 crc kubenswrapper[4930]: I0313 09:36:12.582362 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-848cf88cfc-h7xpv" podUID="341ae9b2-e088-46e1-8a17-60da5410bff8" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.208:5353: i/o timeout" Mar 13 09:36:13 crc kubenswrapper[4930]: I0313 09:36:13.035084 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:36:13 crc kubenswrapper[4930]: I0313 09:36:13.038978 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:36:13 crc kubenswrapper[4930]: I0313 09:36:13.044603 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-57c8b65b96-szpn5" Mar 13 09:36:13 crc kubenswrapper[4930]: I0313 09:36:13.141944 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 09:36:13 crc kubenswrapper[4930]: I0313 09:36:13.156588 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6fd55ffb6d-npc7j"] Mar 13 09:36:13 crc kubenswrapper[4930]: I0313 09:36:13.156734 4930 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 09:36:13 crc kubenswrapper[4930]: I0313 09:36:13.156862 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-6fd55ffb6d-npc7j" podUID="afc45a41-5494-41d9-90af-67a584dac2eb" containerName="placement-log" containerID="cri-o://522d784a063767088689e8af88000689c45e83076a50850a7c9ca1b45432b759" gracePeriod=30 Mar 13 09:36:13 crc kubenswrapper[4930]: I0313 09:36:13.157019 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-6fd55ffb6d-npc7j" podUID="afc45a41-5494-41d9-90af-67a584dac2eb" containerName="placement-api" containerID="cri-o://21b7c4ee405bb6bfc241dca312c28f6dd147ec3b084d0a234ffe5583877eb769" gracePeriod=30 Mar 13 09:36:13 crc kubenswrapper[4930]: I0313 09:36:13.166841 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-6fd55ffb6d-npc7j" podUID="afc45a41-5494-41d9-90af-67a584dac2eb" containerName="placement-log" probeResult="failure" output="Get \"https://10.217.0.203:8778/\": EOF" Mar 13 09:36:13 crc kubenswrapper[4930]: I0313 09:36:13.203126 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 09:36:13 crc kubenswrapper[4930]: I0313 09:36:13.449706 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.215:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 09:36:14 crc kubenswrapper[4930]: I0313 09:36:14.137691 4930 generic.go:334] "Generic (PLEG): container finished" podID="afc45a41-5494-41d9-90af-67a584dac2eb" containerID="522d784a063767088689e8af88000689c45e83076a50850a7c9ca1b45432b759" exitCode=143 Mar 13 09:36:14 crc kubenswrapper[4930]: I0313 09:36:14.138450 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" containerName="cinder-scheduler" containerID="cri-o://b35c75ca056b85198e0f020035b73c2431be9e9688290e0263c297ecee19416b" gracePeriod=30 Mar 13 09:36:14 crc kubenswrapper[4930]: I0313 09:36:14.137746 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fd55ffb6d-npc7j" event={"ID":"afc45a41-5494-41d9-90af-67a584dac2eb","Type":"ContainerDied","Data":"522d784a063767088689e8af88000689c45e83076a50850a7c9ca1b45432b759"} Mar 13 09:36:14 crc kubenswrapper[4930]: I0313 09:36:14.138785 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" containerName="probe" containerID="cri-o://83025acec25b4f21dca183700c7fccc9b6dd4bdf036942fbcf3ca67e0246932c" gracePeriod=30 Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.154482 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-c7549c7dc-lw9qh" Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.162939 4930 generic.go:334] "Generic (PLEG): container finished" podID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" containerID="83025acec25b4f21dca183700c7fccc9b6dd4bdf036942fbcf3ca67e0246932c" exitCode=0 Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.163018 4930 generic.go:334] "Generic (PLEG): container finished" podID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" containerID="b35c75ca056b85198e0f020035b73c2431be9e9688290e0263c297ecee19416b" exitCode=0 Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.163098 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b","Type":"ContainerDied","Data":"83025acec25b4f21dca183700c7fccc9b6dd4bdf036942fbcf3ca67e0246932c"} Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.163233 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b","Type":"ContainerDied","Data":"b35c75ca056b85198e0f020035b73c2431be9e9688290e0263c297ecee19416b"} Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.166921 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1a220dc-49c4-4eb4-b05a-a07f3263b119","Type":"ContainerStarted","Data":"9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c"} Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.167064 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.169847 4930 generic.go:334] "Generic (PLEG): container finished" podID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerID="220750210a0a051baa308d129f907fef78ef4362348c8d034a8333034eaadcfc" exitCode=0 Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.169876 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g6hb8" event={"ID":"2b4f2c08-b04f-40e6-a90c-256225efe861","Type":"ContainerDied","Data":"220750210a0a051baa308d129f907fef78ef4362348c8d034a8333034eaadcfc"} Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.172704 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556576-kctpg" event={"ID":"f4ac8095-d474-4fe0-ab66-5e119896d4a1","Type":"ContainerStarted","Data":"acbe3f948a6eb2d09968373098c52055bb11944d61464e5bbede99904f8b601d"} Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.177528 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bsl74" event={"ID":"a73c1bd6-dd06-4e14-b8ee-1916942e53f5","Type":"ContainerStarted","Data":"1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9"} Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.250007 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.250045 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.269329 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.7080974060000003 podStartE2EDuration="20.269310778s" podCreationTimestamp="2026-03-13 09:35:56 +0000 UTC" firstStartedPulling="2026-03-13 09:35:57.729050428 +0000 UTC m=+1398.478965105" lastFinishedPulling="2026-03-13 09:36:15.2902638 +0000 UTC m=+1416.040178477" observedRunningTime="2026-03-13 09:36:16.264087442 +0000 UTC m=+1417.014002109" watchObservedRunningTime="2026-03-13 09:36:16.269310778 +0000 UTC m=+1417.019225455" Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.273900 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556576-kctpg" podStartSLOduration=3.947955149 podStartE2EDuration="16.273883438s" podCreationTimestamp="2026-03-13 09:36:00 +0000 UTC" firstStartedPulling="2026-03-13 09:36:01.15545463 +0000 UTC m=+1401.905369307" lastFinishedPulling="2026-03-13 09:36:13.481382929 +0000 UTC m=+1414.231297596" observedRunningTime="2026-03-13 09:36:16.241340046 +0000 UTC m=+1416.991254723" watchObservedRunningTime="2026-03-13 09:36:16.273883438 +0000 UTC m=+1417.023798115" Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.289045 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bsl74" podStartSLOduration=5.6058959040000005 podStartE2EDuration="21.289027924s" podCreationTimestamp="2026-03-13 09:35:55 +0000 UTC" firstStartedPulling="2026-03-13 09:35:57.798739082 +0000 UTC m=+1398.548653759" lastFinishedPulling="2026-03-13 09:36:13.481871102 +0000 UTC m=+1414.231785779" observedRunningTime="2026-03-13 09:36:16.288796418 +0000 UTC m=+1417.038711095" watchObservedRunningTime="2026-03-13 09:36:16.289027924 +0000 UTC m=+1417.038942601" Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.800484 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.938068 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-etc-machine-id\") pod \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.938128 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-config-data\") pod \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.938190 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" (UID: "c21aadbe-c137-4bd0-b07a-a114c1e6dc7b"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.938276 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-config-data-custom\") pod \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.938383 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-scripts\") pod \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.938554 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnswt\" (UniqueName: \"kubernetes.io/projected/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-kube-api-access-pnswt\") pod \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.938619 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-combined-ca-bundle\") pod \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\" (UID: \"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b\") " Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.939093 4930 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.945208 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-scripts" (OuterVolumeSpecName: "scripts") pod "c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" (UID: "c21aadbe-c137-4bd0-b07a-a114c1e6dc7b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.947588 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" (UID: "c21aadbe-c137-4bd0-b07a-a114c1e6dc7b"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:16 crc kubenswrapper[4930]: I0313 09:36:16.962697 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-kube-api-access-pnswt" (OuterVolumeSpecName: "kube-api-access-pnswt") pod "c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" (UID: "c21aadbe-c137-4bd0-b07a-a114c1e6dc7b"). InnerVolumeSpecName "kube-api-access-pnswt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.042218 4930 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.042250 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.042261 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnswt\" (UniqueName: \"kubernetes.io/projected/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-kube-api-access-pnswt\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.055036 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" (UID: "c21aadbe-c137-4bd0-b07a-a114c1e6dc7b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.145294 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.153681 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-config-data" (OuterVolumeSpecName: "config-data") pod "c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" (UID: "c21aadbe-c137-4bd0-b07a-a114c1e6dc7b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.190749 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"c21aadbe-c137-4bd0-b07a-a114c1e6dc7b","Type":"ContainerDied","Data":"795e2d664423128de17093f5ab139b5579d4ea9629f5b3d24fb35e27040ae23a"} Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.190824 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.190880 4930 scope.go:117] "RemoveContainer" containerID="83025acec25b4f21dca183700c7fccc9b6dd4bdf036942fbcf3ca67e0246932c" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.199270 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g6hb8" event={"ID":"2b4f2c08-b04f-40e6-a90c-256225efe861","Type":"ContainerStarted","Data":"da6c0c379d767f63fba70506d469a63802e501f36556a8df4f7036d5281ac4a1"} Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.223087 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-g6hb8" podStartSLOduration=3.958904476 podStartE2EDuration="26.223067244s" podCreationTimestamp="2026-03-13 09:35:51 +0000 UTC" firstStartedPulling="2026-03-13 09:35:54.484658551 +0000 UTC m=+1395.234573228" lastFinishedPulling="2026-03-13 09:36:16.748821319 +0000 UTC m=+1417.498735996" observedRunningTime="2026-03-13 09:36:17.216315437 +0000 UTC m=+1417.966230114" watchObservedRunningTime="2026-03-13 09:36:17.223067244 +0000 UTC m=+1417.972981921" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.223646 4930 scope.go:117] "RemoveContainer" containerID="b35c75ca056b85198e0f020035b73c2431be9e9688290e0263c297ecee19416b" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.248332 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.249089 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.264523 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.274331 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 09:36:17 crc kubenswrapper[4930]: E0313 09:36:17.274827 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" containerName="probe" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.274848 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" containerName="probe" Mar 13 09:36:17 crc kubenswrapper[4930]: E0313 09:36:17.274866 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341ae9b2-e088-46e1-8a17-60da5410bff8" containerName="init" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.274872 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="341ae9b2-e088-46e1-8a17-60da5410bff8" containerName="init" Mar 13 09:36:17 crc kubenswrapper[4930]: E0313 09:36:17.274897 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="341ae9b2-e088-46e1-8a17-60da5410bff8" containerName="dnsmasq-dns" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.274904 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="341ae9b2-e088-46e1-8a17-60da5410bff8" containerName="dnsmasq-dns" Mar 13 09:36:17 crc kubenswrapper[4930]: E0313 09:36:17.274923 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" containerName="cinder-scheduler" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.274929 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" containerName="cinder-scheduler" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.275130 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="341ae9b2-e088-46e1-8a17-60da5410bff8" containerName="dnsmasq-dns" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.275151 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" containerName="cinder-scheduler" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.275164 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" containerName="probe" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.276300 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.278653 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.292656 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.341352 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bsl74" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerName="registry-server" probeResult="failure" output=< Mar 13 09:36:17 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:36:17 crc kubenswrapper[4930]: > Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.443255 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.455167 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7365e082-8383-4b59-974c-dfdedb789f77-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.455268 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7365e082-8383-4b59-974c-dfdedb789f77-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.455485 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7365e082-8383-4b59-974c-dfdedb789f77-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.455526 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7365e082-8383-4b59-974c-dfdedb789f77-scripts\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.455739 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7365e082-8383-4b59-974c-dfdedb789f77-config-data\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.455887 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdsgc\" (UniqueName: \"kubernetes.io/projected/7365e082-8383-4b59-974c-dfdedb789f77-kube-api-access-rdsgc\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.557488 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7365e082-8383-4b59-974c-dfdedb789f77-config-data\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.557551 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdsgc\" (UniqueName: \"kubernetes.io/projected/7365e082-8383-4b59-974c-dfdedb789f77-kube-api-access-rdsgc\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.557642 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7365e082-8383-4b59-974c-dfdedb789f77-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.557678 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7365e082-8383-4b59-974c-dfdedb789f77-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.557760 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7365e082-8383-4b59-974c-dfdedb789f77-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.557779 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7365e082-8383-4b59-974c-dfdedb789f77-scripts\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.557782 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7365e082-8383-4b59-974c-dfdedb789f77-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.563209 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7365e082-8383-4b59-974c-dfdedb789f77-scripts\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.563297 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7365e082-8383-4b59-974c-dfdedb789f77-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.564191 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7365e082-8383-4b59-974c-dfdedb789f77-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.567593 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7365e082-8383-4b59-974c-dfdedb789f77-config-data\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.579364 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdsgc\" (UniqueName: \"kubernetes.io/projected/7365e082-8383-4b59-974c-dfdedb789f77-kube-api-access-rdsgc\") pod \"cinder-scheduler-0\" (UID: \"7365e082-8383-4b59-974c-dfdedb789f77\") " pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.596043 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.673338 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/placement-6fd55ffb6d-npc7j" podUID="afc45a41-5494-41d9-90af-67a584dac2eb" containerName="placement-log" probeResult="failure" output="Get \"https://10.217.0.203:8778/\": read tcp 10.217.0.2:54964->10.217.0.203:8778: read: connection reset by peer" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.831837 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.836037 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.840815 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.840929 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.840991 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-qt9x5" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.890788 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.978874 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/234d312d-dc21-441c-bdec-df8606d47c6b-openstack-config-secret\") pod \"openstackclient\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " pod="openstack/openstackclient" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.978965 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234d312d-dc21-441c-bdec-df8606d47c6b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " pod="openstack/openstackclient" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.978995 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrfmb\" (UniqueName: \"kubernetes.io/projected/234d312d-dc21-441c-bdec-df8606d47c6b-kube-api-access-zrfmb\") pod \"openstackclient\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " pod="openstack/openstackclient" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.979050 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/234d312d-dc21-441c-bdec-df8606d47c6b-openstack-config\") pod \"openstackclient\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " pod="openstack/openstackclient" Mar 13 09:36:17 crc kubenswrapper[4930]: I0313 09:36:17.997024 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c21aadbe-c137-4bd0-b07a-a114c1e6dc7b" path="/var/lib/kubelet/pods/c21aadbe-c137-4bd0-b07a-a114c1e6dc7b/volumes" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.083516 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/234d312d-dc21-441c-bdec-df8606d47c6b-openstack-config\") pod \"openstackclient\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.083757 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/234d312d-dc21-441c-bdec-df8606d47c6b-openstack-config-secret\") pod \"openstackclient\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.083787 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234d312d-dc21-441c-bdec-df8606d47c6b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.083824 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrfmb\" (UniqueName: \"kubernetes.io/projected/234d312d-dc21-441c-bdec-df8606d47c6b-kube-api-access-zrfmb\") pod \"openstackclient\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.084758 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/234d312d-dc21-441c-bdec-df8606d47c6b-openstack-config\") pod \"openstackclient\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.089980 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/234d312d-dc21-441c-bdec-df8606d47c6b-openstack-config-secret\") pod \"openstackclient\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.096511 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234d312d-dc21-441c-bdec-df8606d47c6b-combined-ca-bundle\") pod \"openstackclient\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.117184 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrfmb\" (UniqueName: \"kubernetes.io/projected/234d312d-dc21-441c-bdec-df8606d47c6b-kube-api-access-zrfmb\") pod \"openstackclient\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.185987 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.205504 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.252848 4930 generic.go:334] "Generic (PLEG): container finished" podID="f4ac8095-d474-4fe0-ab66-5e119896d4a1" containerID="acbe3f948a6eb2d09968373098c52055bb11944d61464e5bbede99904f8b601d" exitCode=0 Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.252932 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556576-kctpg" event={"ID":"f4ac8095-d474-4fe0-ab66-5e119896d4a1","Type":"ContainerDied","Data":"acbe3f948a6eb2d09968373098c52055bb11944d61464e5bbede99904f8b601d"} Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.257239 4930 generic.go:334] "Generic (PLEG): container finished" podID="afc45a41-5494-41d9-90af-67a584dac2eb" containerID="21b7c4ee405bb6bfc241dca312c28f6dd147ec3b084d0a234ffe5583877eb769" exitCode=0 Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.257356 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fd55ffb6d-npc7j" event={"ID":"afc45a41-5494-41d9-90af-67a584dac2eb","Type":"ContainerDied","Data":"21b7c4ee405bb6bfc241dca312c28f6dd147ec3b084d0a234ffe5583877eb769"} Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.344747 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/openstackclient"] Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.361646 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/openstackclient"] Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.364711 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.385500 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 13 09:36:18 crc kubenswrapper[4930]: E0313 09:36:18.385983 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc45a41-5494-41d9-90af-67a584dac2eb" containerName="placement-api" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.386000 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc45a41-5494-41d9-90af-67a584dac2eb" containerName="placement-api" Mar 13 09:36:18 crc kubenswrapper[4930]: E0313 09:36:18.386015 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afc45a41-5494-41d9-90af-67a584dac2eb" containerName="placement-log" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.386023 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="afc45a41-5494-41d9-90af-67a584dac2eb" containerName="placement-log" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.386256 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc45a41-5494-41d9-90af-67a584dac2eb" containerName="placement-log" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.386282 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="afc45a41-5494-41d9-90af-67a584dac2eb" containerName="placement-api" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.388810 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.401543 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 09:36:18 crc kubenswrapper[4930]: E0313 09:36:18.490443 4930 log.go:32] "RunPodSandbox from runtime service failed" err=< Mar 13 09:36:18 crc kubenswrapper[4930]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_234d312d-dc21-441c-bdec-df8606d47c6b_0(fd5d580e4a0fd881acc9ef687f3d824968fcdfba12a8e4c408243bd24508e11f): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"fd5d580e4a0fd881acc9ef687f3d824968fcdfba12a8e4c408243bd24508e11f" Netns:"/var/run/netns/f9ebca87-4815-4ccf-b9b3-ba388f902a04" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=fd5d580e4a0fd881acc9ef687f3d824968fcdfba12a8e4c408243bd24508e11f;K8S_POD_UID=234d312d-dc21-441c-bdec-df8606d47c6b" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/234d312d-dc21-441c-bdec-df8606d47c6b]: expected pod UID "234d312d-dc21-441c-bdec-df8606d47c6b" but got "9b1d8908-3b46-4270-b04c-6ad122632782" from Kube API Mar 13 09:36:18 crc kubenswrapper[4930]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 09:36:18 crc kubenswrapper[4930]: > Mar 13 09:36:18 crc kubenswrapper[4930]: E0313 09:36:18.490555 4930 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err=< Mar 13 09:36:18 crc kubenswrapper[4930]: rpc error: code = Unknown desc = failed to create pod network sandbox k8s_openstackclient_openstack_234d312d-dc21-441c-bdec-df8606d47c6b_0(fd5d580e4a0fd881acc9ef687f3d824968fcdfba12a8e4c408243bd24508e11f): error adding pod openstack_openstackclient to CNI network "multus-cni-network": plugin type="multus-shim" name="multus-cni-network" failed (add): CmdAdd (shim): CNI request failed with status 400: 'ContainerID:"fd5d580e4a0fd881acc9ef687f3d824968fcdfba12a8e4c408243bd24508e11f" Netns:"/var/run/netns/f9ebca87-4815-4ccf-b9b3-ba388f902a04" IfName:"eth0" Args:"IgnoreUnknown=1;K8S_POD_NAMESPACE=openstack;K8S_POD_NAME=openstackclient;K8S_POD_INFRA_CONTAINER_ID=fd5d580e4a0fd881acc9ef687f3d824968fcdfba12a8e4c408243bd24508e11f;K8S_POD_UID=234d312d-dc21-441c-bdec-df8606d47c6b" Path:"" ERRORED: error configuring pod [openstack/openstackclient] networking: Multus: [openstack/openstackclient/234d312d-dc21-441c-bdec-df8606d47c6b]: expected pod UID "234d312d-dc21-441c-bdec-df8606d47c6b" but got "9b1d8908-3b46-4270-b04c-6ad122632782" from Kube API Mar 13 09:36:18 crc kubenswrapper[4930]: ': StdinData: {"binDir":"/var/lib/cni/bin","clusterNetwork":"/host/run/multus/cni/net.d/10-ovn-kubernetes.conf","cniVersion":"0.3.1","daemonSocketDir":"/run/multus/socket","globalNamespaces":"default,openshift-multus,openshift-sriov-network-operator,openshift-cnv","logLevel":"verbose","logToStderr":true,"name":"multus-cni-network","namespaceIsolation":true,"type":"multus-shim"} Mar 13 09:36:18 crc kubenswrapper[4930]: > pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.498204 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vl5sz\" (UniqueName: \"kubernetes.io/projected/afc45a41-5494-41d9-90af-67a584dac2eb-kube-api-access-vl5sz\") pod \"afc45a41-5494-41d9-90af-67a584dac2eb\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.498342 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afc45a41-5494-41d9-90af-67a584dac2eb-logs\") pod \"afc45a41-5494-41d9-90af-67a584dac2eb\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.498461 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-config-data\") pod \"afc45a41-5494-41d9-90af-67a584dac2eb\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.498570 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-internal-tls-certs\") pod \"afc45a41-5494-41d9-90af-67a584dac2eb\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.498607 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-combined-ca-bundle\") pod \"afc45a41-5494-41d9-90af-67a584dac2eb\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.498633 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-scripts\") pod \"afc45a41-5494-41d9-90af-67a584dac2eb\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.498652 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-public-tls-certs\") pod \"afc45a41-5494-41d9-90af-67a584dac2eb\" (UID: \"afc45a41-5494-41d9-90af-67a584dac2eb\") " Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.498964 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs76c\" (UniqueName: \"kubernetes.io/projected/9b1d8908-3b46-4270-b04c-6ad122632782-kube-api-access-gs76c\") pod \"openstackclient\" (UID: \"9b1d8908-3b46-4270-b04c-6ad122632782\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.499012 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9b1d8908-3b46-4270-b04c-6ad122632782-openstack-config-secret\") pod \"openstackclient\" (UID: \"9b1d8908-3b46-4270-b04c-6ad122632782\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.499043 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b1d8908-3b46-4270-b04c-6ad122632782-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9b1d8908-3b46-4270-b04c-6ad122632782\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.499062 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9b1d8908-3b46-4270-b04c-6ad122632782-openstack-config\") pod \"openstackclient\" (UID: \"9b1d8908-3b46-4270-b04c-6ad122632782\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.499092 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afc45a41-5494-41d9-90af-67a584dac2eb-logs" (OuterVolumeSpecName: "logs") pod "afc45a41-5494-41d9-90af-67a584dac2eb" (UID: "afc45a41-5494-41d9-90af-67a584dac2eb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.499551 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/afc45a41-5494-41d9-90af-67a584dac2eb-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.505790 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afc45a41-5494-41d9-90af-67a584dac2eb-kube-api-access-vl5sz" (OuterVolumeSpecName: "kube-api-access-vl5sz") pod "afc45a41-5494-41d9-90af-67a584dac2eb" (UID: "afc45a41-5494-41d9-90af-67a584dac2eb"). InnerVolumeSpecName "kube-api-access-vl5sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.523614 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-scripts" (OuterVolumeSpecName: "scripts") pod "afc45a41-5494-41d9-90af-67a584dac2eb" (UID: "afc45a41-5494-41d9-90af-67a584dac2eb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.574568 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "afc45a41-5494-41d9-90af-67a584dac2eb" (UID: "afc45a41-5494-41d9-90af-67a584dac2eb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.601679 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs76c\" (UniqueName: \"kubernetes.io/projected/9b1d8908-3b46-4270-b04c-6ad122632782-kube-api-access-gs76c\") pod \"openstackclient\" (UID: \"9b1d8908-3b46-4270-b04c-6ad122632782\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.601744 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9b1d8908-3b46-4270-b04c-6ad122632782-openstack-config-secret\") pod \"openstackclient\" (UID: \"9b1d8908-3b46-4270-b04c-6ad122632782\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.601819 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b1d8908-3b46-4270-b04c-6ad122632782-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9b1d8908-3b46-4270-b04c-6ad122632782\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.601847 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9b1d8908-3b46-4270-b04c-6ad122632782-openstack-config\") pod \"openstackclient\" (UID: \"9b1d8908-3b46-4270-b04c-6ad122632782\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.602243 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vl5sz\" (UniqueName: \"kubernetes.io/projected/afc45a41-5494-41d9-90af-67a584dac2eb-kube-api-access-vl5sz\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.602262 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.602272 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.605999 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/9b1d8908-3b46-4270-b04c-6ad122632782-openstack-config\") pod \"openstackclient\" (UID: \"9b1d8908-3b46-4270-b04c-6ad122632782\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.609047 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9b1d8908-3b46-4270-b04c-6ad122632782-combined-ca-bundle\") pod \"openstackclient\" (UID: \"9b1d8908-3b46-4270-b04c-6ad122632782\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.609697 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/9b1d8908-3b46-4270-b04c-6ad122632782-openstack-config-secret\") pod \"openstackclient\" (UID: \"9b1d8908-3b46-4270-b04c-6ad122632782\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.631057 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-config-data" (OuterVolumeSpecName: "config-data") pod "afc45a41-5494-41d9-90af-67a584dac2eb" (UID: "afc45a41-5494-41d9-90af-67a584dac2eb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.639930 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs76c\" (UniqueName: \"kubernetes.io/projected/9b1d8908-3b46-4270-b04c-6ad122632782-kube-api-access-gs76c\") pod \"openstackclient\" (UID: \"9b1d8908-3b46-4270-b04c-6ad122632782\") " pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.703002 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "afc45a41-5494-41d9-90af-67a584dac2eb" (UID: "afc45a41-5494-41d9-90af-67a584dac2eb"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.704685 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.704714 4930 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.713576 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.748943 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "afc45a41-5494-41d9-90af-67a584dac2eb" (UID: "afc45a41-5494-41d9-90af-67a584dac2eb"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:18 crc kubenswrapper[4930]: I0313 09:36:18.816920 4930 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/afc45a41-5494-41d9-90af-67a584dac2eb-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.306604 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6fd55ffb6d-npc7j" event={"ID":"afc45a41-5494-41d9-90af-67a584dac2eb","Type":"ContainerDied","Data":"470b76b6d6bf3e77ed12b3aaac0a9c61e8f2e2775e634961299f7f36bedc3323"} Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.307080 4930 scope.go:117] "RemoveContainer" containerID="21b7c4ee405bb6bfc241dca312c28f6dd147ec3b084d0a234ffe5583877eb769" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.307273 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6fd55ffb6d-npc7j" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.313107 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7365e082-8383-4b59-974c-dfdedb789f77","Type":"ContainerStarted","Data":"7427f4f5722b28f8d487b62b99ea7f5be716124c600a214b0b1d7690a9e1c14d"} Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.313143 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7365e082-8383-4b59-974c-dfdedb789f77","Type":"ContainerStarted","Data":"e4972853f07852d895d5248aebbf92e7c44b1e2d281eb9d72be282d95aa1a7b6"} Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.313174 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.330021 4930 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="234d312d-dc21-441c-bdec-df8606d47c6b" podUID="9b1d8908-3b46-4270-b04c-6ad122632782" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.346512 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.373187 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.396622 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6fd55ffb6d-npc7j"] Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.411539 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6fd55ffb6d-npc7j"] Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.453585 4930 scope.go:117] "RemoveContainer" containerID="522d784a063767088689e8af88000689c45e83076a50850a7c9ca1b45432b759" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.545319 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrfmb\" (UniqueName: \"kubernetes.io/projected/234d312d-dc21-441c-bdec-df8606d47c6b-kube-api-access-zrfmb\") pod \"234d312d-dc21-441c-bdec-df8606d47c6b\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.545528 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/234d312d-dc21-441c-bdec-df8606d47c6b-openstack-config-secret\") pod \"234d312d-dc21-441c-bdec-df8606d47c6b\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.545732 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234d312d-dc21-441c-bdec-df8606d47c6b-combined-ca-bundle\") pod \"234d312d-dc21-441c-bdec-df8606d47c6b\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.545799 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/234d312d-dc21-441c-bdec-df8606d47c6b-openstack-config\") pod \"234d312d-dc21-441c-bdec-df8606d47c6b\" (UID: \"234d312d-dc21-441c-bdec-df8606d47c6b\") " Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.546965 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/234d312d-dc21-441c-bdec-df8606d47c6b-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "234d312d-dc21-441c-bdec-df8606d47c6b" (UID: "234d312d-dc21-441c-bdec-df8606d47c6b"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.554418 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234d312d-dc21-441c-bdec-df8606d47c6b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "234d312d-dc21-441c-bdec-df8606d47c6b" (UID: "234d312d-dc21-441c-bdec-df8606d47c6b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.556987 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/234d312d-dc21-441c-bdec-df8606d47c6b-kube-api-access-zrfmb" (OuterVolumeSpecName: "kube-api-access-zrfmb") pod "234d312d-dc21-441c-bdec-df8606d47c6b" (UID: "234d312d-dc21-441c-bdec-df8606d47c6b"). InnerVolumeSpecName "kube-api-access-zrfmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.564557 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/234d312d-dc21-441c-bdec-df8606d47c6b-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "234d312d-dc21-441c-bdec-df8606d47c6b" (UID: "234d312d-dc21-441c-bdec-df8606d47c6b"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.648158 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/234d312d-dc21-441c-bdec-df8606d47c6b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.648194 4930 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/234d312d-dc21-441c-bdec-df8606d47c6b-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.648208 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrfmb\" (UniqueName: \"kubernetes.io/projected/234d312d-dc21-441c-bdec-df8606d47c6b-kube-api-access-zrfmb\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:19 crc kubenswrapper[4930]: I0313 09:36:19.648221 4930 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/234d312d-dc21-441c-bdec-df8606d47c6b-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.001394 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="234d312d-dc21-441c-bdec-df8606d47c6b" path="/var/lib/kubelet/pods/234d312d-dc21-441c-bdec-df8606d47c6b/volumes" Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.002808 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afc45a41-5494-41d9-90af-67a584dac2eb" path="/var/lib/kubelet/pods/afc45a41-5494-41d9-90af-67a584dac2eb/volumes" Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.194834 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556576-kctpg" Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.347328 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9b1d8908-3b46-4270-b04c-6ad122632782","Type":"ContainerStarted","Data":"4e3dbb9a79c4770924790839319dfba2899f8de1af73075cdf0391b991248747"} Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.349510 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556576-kctpg" event={"ID":"f4ac8095-d474-4fe0-ab66-5e119896d4a1","Type":"ContainerDied","Data":"0805b88195b489aba14cc8e951582338c1d17735d863808e1cb92dbbefa47e38"} Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.349563 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0805b88195b489aba14cc8e951582338c1d17735d863808e1cb92dbbefa47e38" Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.349578 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556576-kctpg" Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.377557 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.378962 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7365e082-8383-4b59-974c-dfdedb789f77","Type":"ContainerStarted","Data":"7ac6c3976584c65f2c464bea07b243083e16098847f8826919d78413969f9d16"} Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.384192 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8qcj5\" (UniqueName: \"kubernetes.io/projected/f4ac8095-d474-4fe0-ab66-5e119896d4a1-kube-api-access-8qcj5\") pod \"f4ac8095-d474-4fe0-ab66-5e119896d4a1\" (UID: \"f4ac8095-d474-4fe0-ab66-5e119896d4a1\") " Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.391910 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f4ac8095-d474-4fe0-ab66-5e119896d4a1-kube-api-access-8qcj5" (OuterVolumeSpecName: "kube-api-access-8qcj5") pod "f4ac8095-d474-4fe0-ab66-5e119896d4a1" (UID: "f4ac8095-d474-4fe0-ab66-5e119896d4a1"). InnerVolumeSpecName "kube-api-access-8qcj5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.394037 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556570-qjp4p"] Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.410285 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556570-qjp4p"] Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.431245 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.431227992 podStartE2EDuration="3.431227992s" podCreationTimestamp="2026-03-13 09:36:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:20.425478102 +0000 UTC m=+1421.175392779" watchObservedRunningTime="2026-03-13 09:36:20.431227992 +0000 UTC m=+1421.181142659" Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.431279 4930 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openstack/openstackclient" oldPodUID="234d312d-dc21-441c-bdec-df8606d47c6b" podUID="9b1d8908-3b46-4270-b04c-6ad122632782" Mar 13 09:36:20 crc kubenswrapper[4930]: I0313 09:36:20.490077 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8qcj5\" (UniqueName: \"kubernetes.io/projected/f4ac8095-d474-4fe0-ab66-5e119896d4a1-kube-api-access-8qcj5\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.177741 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-62v7d"] Mar 13 09:36:21 crc kubenswrapper[4930]: E0313 09:36:21.178517 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4ac8095-d474-4fe0-ab66-5e119896d4a1" containerName="oc" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.178530 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4ac8095-d474-4fe0-ab66-5e119896d4a1" containerName="oc" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.179527 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4ac8095-d474-4fe0-ab66-5e119896d4a1" containerName="oc" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.181092 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.205396 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-62v7d"] Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.308533 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7sw8\" (UniqueName: \"kubernetes.io/projected/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-kube-api-access-j7sw8\") pod \"redhat-marketplace-62v7d\" (UID: \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\") " pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.308700 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-catalog-content\") pod \"redhat-marketplace-62v7d\" (UID: \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\") " pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.309038 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-utilities\") pod \"redhat-marketplace-62v7d\" (UID: \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\") " pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.413124 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-catalog-content\") pod \"redhat-marketplace-62v7d\" (UID: \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\") " pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.413282 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-utilities\") pod \"redhat-marketplace-62v7d\" (UID: \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\") " pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.413370 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7sw8\" (UniqueName: \"kubernetes.io/projected/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-kube-api-access-j7sw8\") pod \"redhat-marketplace-62v7d\" (UID: \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\") " pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.413801 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-catalog-content\") pod \"redhat-marketplace-62v7d\" (UID: \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\") " pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.414049 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-utilities\") pod \"redhat-marketplace-62v7d\" (UID: \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\") " pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.438358 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7sw8\" (UniqueName: \"kubernetes.io/projected/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-kube-api-access-j7sw8\") pod \"redhat-marketplace-62v7d\" (UID: \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\") " pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.505362 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:21 crc kubenswrapper[4930]: I0313 09:36:21.992291 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c44fc59-e3a9-448a-b593-d572109a37c4" path="/var/lib/kubelet/pods/1c44fc59-e3a9-448a-b593-d572109a37c4/volumes" Mar 13 09:36:22 crc kubenswrapper[4930]: I0313 09:36:22.081422 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-62v7d"] Mar 13 09:36:22 crc kubenswrapper[4930]: I0313 09:36:22.238905 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:36:22 crc kubenswrapper[4930]: I0313 09:36:22.239227 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:36:22 crc kubenswrapper[4930]: I0313 09:36:22.300110 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:36:22 crc kubenswrapper[4930]: I0313 09:36:22.403336 4930 generic.go:334] "Generic (PLEG): container finished" podID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" containerID="6028c0d550eee00bf31309724244e206fe8871eee28e65a2e6ab2933370f8753" exitCode=0 Mar 13 09:36:22 crc kubenswrapper[4930]: I0313 09:36:22.403556 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62v7d" event={"ID":"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929","Type":"ContainerDied","Data":"6028c0d550eee00bf31309724244e206fe8871eee28e65a2e6ab2933370f8753"} Mar 13 09:36:22 crc kubenswrapper[4930]: I0313 09:36:22.403622 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62v7d" event={"ID":"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929","Type":"ContainerStarted","Data":"f6d2a71e9f06179a3a119543c8e02d642bd682240e2c1907d9a0b15f1d3b8545"} Mar 13 09:36:22 crc kubenswrapper[4930]: I0313 09:36:22.473576 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-g6hb8" Mar 13 09:36:22 crc kubenswrapper[4930]: I0313 09:36:22.597045 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 09:36:22 crc kubenswrapper[4930]: I0313 09:36:22.858484 4930 scope.go:117] "RemoveContainer" containerID="ede3f1783e9d36de90ca201f79679248e68b7f2d10a2509424618bd62ed0b6fa" Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.216606 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5cc4fbb59f-99vq6" Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.285767 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6dcf4d8cb8-7d6r6"] Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.286200 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6dcf4d8cb8-7d6r6" podUID="aef74c6a-403d-4388-a873-43dca210f617" containerName="neutron-api" containerID="cri-o://93155ac78aa8e7896ad1c995dc5aac22c196b3c9ecdcd86f67cbd7bfb454b789" gracePeriod=30 Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.286648 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6dcf4d8cb8-7d6r6" podUID="aef74c6a-403d-4388-a873-43dca210f617" containerName="neutron-httpd" containerID="cri-o://35ba88787bfb2c888f48f45e752c3ba2753ebbdbab7fbe30723229f6d66e42fc" gracePeriod=30 Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.814031 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-g6hb8"] Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.898496 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-7f4b8495c9-mqbjk"] Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.900405 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.914364 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.914449 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-tbzqx" Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.914618 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.929930 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7f4b8495c9-mqbjk"] Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.988153 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-config-data\") pod \"heat-engine-7f4b8495c9-mqbjk\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.988220 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pxz6\" (UniqueName: \"kubernetes.io/projected/7754b93c-5665-4f0d-8e26-d9910d1e90c4-kube-api-access-7pxz6\") pod \"heat-engine-7f4b8495c9-mqbjk\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.988273 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-combined-ca-bundle\") pod \"heat-engine-7f4b8495c9-mqbjk\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:23 crc kubenswrapper[4930]: I0313 09:36:23.988460 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-config-data-custom\") pod \"heat-engine-7f4b8495c9-mqbjk\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.011922 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-f4dfb5f5b-pw8zp"] Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.013505 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.022879 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.036964 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-f4dfb5f5b-pw8zp"] Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.066504 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-hr7c4"] Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.068515 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.089749 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-config-data-custom\") pod \"heat-engine-7f4b8495c9-mqbjk\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.089810 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-config-data-custom\") pod \"heat-cfnapi-f4dfb5f5b-pw8zp\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.089861 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tq9tq\" (UniqueName: \"kubernetes.io/projected/7548398c-45a9-4165-81fe-3f15f68dc161-kube-api-access-tq9tq\") pod \"heat-cfnapi-f4dfb5f5b-pw8zp\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.089879 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-combined-ca-bundle\") pod \"heat-cfnapi-f4dfb5f5b-pw8zp\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.089899 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-config-data\") pod \"heat-cfnapi-f4dfb5f5b-pw8zp\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.089930 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-config-data\") pod \"heat-engine-7f4b8495c9-mqbjk\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.089959 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pxz6\" (UniqueName: \"kubernetes.io/projected/7754b93c-5665-4f0d-8e26-d9910d1e90c4-kube-api-access-7pxz6\") pod \"heat-engine-7f4b8495c9-mqbjk\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.089982 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-combined-ca-bundle\") pod \"heat-engine-7f4b8495c9-mqbjk\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.090204 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-hr7c4"] Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.113541 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-config-data-custom\") pod \"heat-engine-7f4b8495c9-mqbjk\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.113958 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-combined-ca-bundle\") pod \"heat-engine-7f4b8495c9-mqbjk\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.116759 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-config-data\") pod \"heat-engine-7f4b8495c9-mqbjk\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.127636 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pxz6\" (UniqueName: \"kubernetes.io/projected/7754b93c-5665-4f0d-8e26-d9910d1e90c4-kube-api-access-7pxz6\") pod \"heat-engine-7f4b8495c9-mqbjk\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.164114 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-558f5d58ff-jkrpd"] Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.165833 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.176859 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.197243 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.197382 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-config\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.197458 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.197587 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqf7d\" (UniqueName: \"kubernetes.io/projected/a2e994f9-023c-4cb7-b6b7-5a974a78e746-kube-api-access-sqf7d\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.197679 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-config-data-custom\") pod \"heat-cfnapi-f4dfb5f5b-pw8zp\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.197742 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.197769 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tq9tq\" (UniqueName: \"kubernetes.io/projected/7548398c-45a9-4165-81fe-3f15f68dc161-kube-api-access-tq9tq\") pod \"heat-cfnapi-f4dfb5f5b-pw8zp\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.197788 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-combined-ca-bundle\") pod \"heat-cfnapi-f4dfb5f5b-pw8zp\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.197808 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-config-data\") pod \"heat-cfnapi-f4dfb5f5b-pw8zp\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.197870 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.210284 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-558f5d58ff-jkrpd"] Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.212053 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-config-data-custom\") pod \"heat-cfnapi-f4dfb5f5b-pw8zp\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.213473 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-combined-ca-bundle\") pod \"heat-cfnapi-f4dfb5f5b-pw8zp\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.232002 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-config-data\") pod \"heat-cfnapi-f4dfb5f5b-pw8zp\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.232697 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tq9tq\" (UniqueName: \"kubernetes.io/projected/7548398c-45a9-4165-81fe-3f15f68dc161-kube-api-access-tq9tq\") pod \"heat-cfnapi-f4dfb5f5b-pw8zp\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.245750 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.262184 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z4n72"] Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.262459 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-z4n72" podUID="918d3d63-22c7-433d-86c8-22747f6d3e77" containerName="registry-server" containerID="cri-o://afac97426e2bd5db823c7655dde77c9560e76c4e36127f86764ec3dc68f6a106" gracePeriod=2 Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.300127 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.300574 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.300623 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.300683 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-config-data\") pod \"heat-api-558f5d58ff-jkrpd\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.300728 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-config\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.300771 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln75n\" (UniqueName: \"kubernetes.io/projected/409ff6c8-9776-400c-9205-58658655a6a2-kube-api-access-ln75n\") pod \"heat-api-558f5d58ff-jkrpd\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.300809 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.300849 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-config-data-custom\") pod \"heat-api-558f5d58ff-jkrpd\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.300907 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-combined-ca-bundle\") pod \"heat-api-558f5d58ff-jkrpd\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.300935 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqf7d\" (UniqueName: \"kubernetes.io/projected/a2e994f9-023c-4cb7-b6b7-5a974a78e746-kube-api-access-sqf7d\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.302292 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-config\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.303306 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-dns-swift-storage-0\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.303795 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-ovsdbserver-nb\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.305807 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-ovsdbserver-sb\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.309837 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-dns-svc\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.325147 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqf7d\" (UniqueName: \"kubernetes.io/projected/a2e994f9-023c-4cb7-b6b7-5a974a78e746-kube-api-access-sqf7d\") pod \"dnsmasq-dns-688b9f5b49-hr7c4\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.388393 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.391143 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.408737 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-config-data\") pod \"heat-api-558f5d58ff-jkrpd\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.408813 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln75n\" (UniqueName: \"kubernetes.io/projected/409ff6c8-9776-400c-9205-58658655a6a2-kube-api-access-ln75n\") pod \"heat-api-558f5d58ff-jkrpd\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.408857 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-config-data-custom\") pod \"heat-api-558f5d58ff-jkrpd\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.408898 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-combined-ca-bundle\") pod \"heat-api-558f5d58ff-jkrpd\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.417034 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-config-data-custom\") pod \"heat-api-558f5d58ff-jkrpd\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.423228 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-combined-ca-bundle\") pod \"heat-api-558f5d58ff-jkrpd\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.439143 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-config-data\") pod \"heat-api-558f5d58ff-jkrpd\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.459093 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln75n\" (UniqueName: \"kubernetes.io/projected/409ff6c8-9776-400c-9205-58658655a6a2-kube-api-access-ln75n\") pod \"heat-api-558f5d58ff-jkrpd\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.540617 4930 generic.go:334] "Generic (PLEG): container finished" podID="aef74c6a-403d-4388-a873-43dca210f617" containerID="35ba88787bfb2c888f48f45e752c3ba2753ebbdbab7fbe30723229f6d66e42fc" exitCode=0 Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.540723 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dcf4d8cb8-7d6r6" event={"ID":"aef74c6a-403d-4388-a873-43dca210f617","Type":"ContainerDied","Data":"35ba88787bfb2c888f48f45e752c3ba2753ebbdbab7fbe30723229f6d66e42fc"} Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.556629 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62v7d" event={"ID":"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929","Type":"ContainerStarted","Data":"d32a14d102f8fdb0a134eac03d3661cbe076acce5ba1643171c5c8f9157b177d"} Mar 13 09:36:24 crc kubenswrapper[4930]: I0313 09:36:24.751866 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.244211 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-7f4b8495c9-mqbjk"] Mar 13 09:36:25 crc kubenswrapper[4930]: W0313 09:36:25.384056 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7548398c_45a9_4165_81fe_3f15f68dc161.slice/crio-c4d8a90d361b52f4759606081f4ed85a048543f130355db7bbb67e82d25c4cf6 WatchSource:0}: Error finding container c4d8a90d361b52f4759606081f4ed85a048543f130355db7bbb67e82d25c4cf6: Status 404 returned error can't find the container with id c4d8a90d361b52f4759606081f4ed85a048543f130355db7bbb67e82d25c4cf6 Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.421849 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-f4dfb5f5b-pw8zp"] Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.443078 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-hr7c4"] Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.589854 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7f4b8495c9-mqbjk" event={"ID":"7754b93c-5665-4f0d-8e26-d9910d1e90c4","Type":"ContainerStarted","Data":"cbf3e3fbe0c39b5610758ce1592b65352077ac1dedf56f0fe4939c1a38009b9e"} Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.651733 4930 generic.go:334] "Generic (PLEG): container finished" podID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" containerID="d32a14d102f8fdb0a134eac03d3661cbe076acce5ba1643171c5c8f9157b177d" exitCode=0 Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.651812 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62v7d" event={"ID":"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929","Type":"ContainerDied","Data":"d32a14d102f8fdb0a134eac03d3661cbe076acce5ba1643171c5c8f9157b177d"} Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.657559 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" event={"ID":"a2e994f9-023c-4cb7-b6b7-5a974a78e746","Type":"ContainerStarted","Data":"ae7e4431da1cc8392755493e685b1b9a59e9329e5af0b0c4b3361f20a3042872"} Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.698682 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" event={"ID":"7548398c-45a9-4165-81fe-3f15f68dc161","Type":"ContainerStarted","Data":"c4d8a90d361b52f4759606081f4ed85a048543f130355db7bbb67e82d25c4cf6"} Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.707286 4930 generic.go:334] "Generic (PLEG): container finished" podID="918d3d63-22c7-433d-86c8-22747f6d3e77" containerID="afac97426e2bd5db823c7655dde77c9560e76c4e36127f86764ec3dc68f6a106" exitCode=0 Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.707339 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4n72" event={"ID":"918d3d63-22c7-433d-86c8-22747f6d3e77","Type":"ContainerDied","Data":"afac97426e2bd5db823c7655dde77c9560e76c4e36127f86764ec3dc68f6a106"} Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.763237 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.803602 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-558f5d58ff-jkrpd"] Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.821959 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/918d3d63-22c7-433d-86c8-22747f6d3e77-catalog-content\") pod \"918d3d63-22c7-433d-86c8-22747f6d3e77\" (UID: \"918d3d63-22c7-433d-86c8-22747f6d3e77\") " Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.822072 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bqtg\" (UniqueName: \"kubernetes.io/projected/918d3d63-22c7-433d-86c8-22747f6d3e77-kube-api-access-4bqtg\") pod \"918d3d63-22c7-433d-86c8-22747f6d3e77\" (UID: \"918d3d63-22c7-433d-86c8-22747f6d3e77\") " Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.829022 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/918d3d63-22c7-433d-86c8-22747f6d3e77-kube-api-access-4bqtg" (OuterVolumeSpecName: "kube-api-access-4bqtg") pod "918d3d63-22c7-433d-86c8-22747f6d3e77" (UID: "918d3d63-22c7-433d-86c8-22747f6d3e77"). InnerVolumeSpecName "kube-api-access-4bqtg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.933585 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/918d3d63-22c7-433d-86c8-22747f6d3e77-utilities\") pod \"918d3d63-22c7-433d-86c8-22747f6d3e77\" (UID: \"918d3d63-22c7-433d-86c8-22747f6d3e77\") " Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.934196 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/918d3d63-22c7-433d-86c8-22747f6d3e77-utilities" (OuterVolumeSpecName: "utilities") pod "918d3d63-22c7-433d-86c8-22747f6d3e77" (UID: "918d3d63-22c7-433d-86c8-22747f6d3e77"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.936226 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/918d3d63-22c7-433d-86c8-22747f6d3e77-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.936451 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bqtg\" (UniqueName: \"kubernetes.io/projected/918d3d63-22c7-433d-86c8-22747f6d3e77-kube-api-access-4bqtg\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:25 crc kubenswrapper[4930]: I0313 09:36:25.975040 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/918d3d63-22c7-433d-86c8-22747f6d3e77-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "918d3d63-22c7-433d-86c8-22747f6d3e77" (UID: "918d3d63-22c7-433d-86c8-22747f6d3e77"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.040818 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/918d3d63-22c7-433d-86c8-22747f6d3e77-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.192879 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-bd7b6bc5f-hcdp4"] Mar 13 09:36:26 crc kubenswrapper[4930]: E0313 09:36:26.193343 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="918d3d63-22c7-433d-86c8-22747f6d3e77" containerName="extract-utilities" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.193360 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="918d3d63-22c7-433d-86c8-22747f6d3e77" containerName="extract-utilities" Mar 13 09:36:26 crc kubenswrapper[4930]: E0313 09:36:26.193388 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="918d3d63-22c7-433d-86c8-22747f6d3e77" containerName="extract-content" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.193417 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="918d3d63-22c7-433d-86c8-22747f6d3e77" containerName="extract-content" Mar 13 09:36:26 crc kubenswrapper[4930]: E0313 09:36:26.193457 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="918d3d63-22c7-433d-86c8-22747f6d3e77" containerName="registry-server" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.193464 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="918d3d63-22c7-433d-86c8-22747f6d3e77" containerName="registry-server" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.193670 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="918d3d63-22c7-433d-86c8-22747f6d3e77" containerName="registry-server" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.194875 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.198023 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.198220 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.198443 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.236354 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-bd7b6bc5f-hcdp4"] Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.244948 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-internal-tls-certs\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.245029 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-config-data\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.245105 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-run-httpd\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.245127 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-log-httpd\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.245148 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-etc-swift\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.245163 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-combined-ca-bundle\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.245240 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk65n\" (UniqueName: \"kubernetes.io/projected/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-kube-api-access-vk65n\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.245273 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-public-tls-certs\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.347618 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-public-tls-certs\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.347700 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-internal-tls-certs\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.347759 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-config-data\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.347837 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-run-httpd\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.347860 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-log-httpd\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.347885 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-etc-swift\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.347901 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-combined-ca-bundle\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.347981 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk65n\" (UniqueName: \"kubernetes.io/projected/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-kube-api-access-vk65n\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.349600 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-run-httpd\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.352085 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-log-httpd\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.355477 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-combined-ca-bundle\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.358499 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-public-tls-certs\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.358551 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-config-data\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.372151 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk65n\" (UniqueName: \"kubernetes.io/projected/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-kube-api-access-vk65n\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.381988 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-etc-swift\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.403096 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b0eff861-dcbe-4a46-97d1-bc1b4558a6ab-internal-tls-certs\") pod \"swift-proxy-bd7b6bc5f-hcdp4\" (UID: \"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab\") " pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.523620 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.762744 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-z4n72" event={"ID":"918d3d63-22c7-433d-86c8-22747f6d3e77","Type":"ContainerDied","Data":"7404131f69d6f13496d7a648edce854bd3b3b6057a056036b371524c1263dfbf"} Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.763110 4930 scope.go:117] "RemoveContainer" containerID="afac97426e2bd5db823c7655dde77c9560e76c4e36127f86764ec3dc68f6a106" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.762777 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-z4n72" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.803874 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7f4b8495c9-mqbjk" event={"ID":"7754b93c-5665-4f0d-8e26-d9910d1e90c4","Type":"ContainerStarted","Data":"db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e"} Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.805661 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.830656 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-558f5d58ff-jkrpd" event={"ID":"409ff6c8-9776-400c-9205-58658655a6a2","Type":"ContainerStarted","Data":"00865374bebca57b67f00c191ddb576709c34011519e45d5222bf35ac1ef1c48"} Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.853865 4930 generic.go:334] "Generic (PLEG): container finished" podID="a2e994f9-023c-4cb7-b6b7-5a974a78e746" containerID="359ba1e318d6db010d5d820cadf70025e50d9954b26f32eeaa8ccf66e62215f2" exitCode=0 Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.853922 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" event={"ID":"a2e994f9-023c-4cb7-b6b7-5a974a78e746","Type":"ContainerDied","Data":"359ba1e318d6db010d5d820cadf70025e50d9954b26f32eeaa8ccf66e62215f2"} Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.865482 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-z4n72"] Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.875747 4930 scope.go:117] "RemoveContainer" containerID="248a5fd5b70b5f6fcfaad1325a8b0cf6771095ef6f4792eae75699a9600dbf21" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.936508 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-z4n72"] Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.944096 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-7f4b8495c9-mqbjk" podStartSLOduration=3.944073985 podStartE2EDuration="3.944073985s" podCreationTimestamp="2026-03-13 09:36:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:26.837831034 +0000 UTC m=+1427.587745711" watchObservedRunningTime="2026-03-13 09:36:26.944073985 +0000 UTC m=+1427.693988662" Mar 13 09:36:26 crc kubenswrapper[4930]: I0313 09:36:26.948772 4930 scope.go:117] "RemoveContainer" containerID="bacc91f65d692a20555d1d03af80b7ec87a0c091750d288483510ec0196a3660" Mar 13 09:36:27 crc kubenswrapper[4930]: I0313 09:36:27.085506 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 09:36:27 crc kubenswrapper[4930]: I0313 09:36:27.315200 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bsl74" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerName="registry-server" probeResult="failure" output=< Mar 13 09:36:27 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:36:27 crc kubenswrapper[4930]: > Mar 13 09:36:27 crc kubenswrapper[4930]: I0313 09:36:27.397314 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-bd7b6bc5f-hcdp4"] Mar 13 09:36:27 crc kubenswrapper[4930]: I0313 09:36:27.966531 4930 generic.go:334] "Generic (PLEG): container finished" podID="aef74c6a-403d-4388-a873-43dca210f617" containerID="93155ac78aa8e7896ad1c995dc5aac22c196b3c9ecdcd86f67cbd7bfb454b789" exitCode=0 Mar 13 09:36:27 crc kubenswrapper[4930]: I0313 09:36:27.967128 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dcf4d8cb8-7d6r6" event={"ID":"aef74c6a-403d-4388-a873-43dca210f617","Type":"ContainerDied","Data":"93155ac78aa8e7896ad1c995dc5aac22c196b3c9ecdcd86f67cbd7bfb454b789"} Mar 13 09:36:28 crc kubenswrapper[4930]: I0313 09:36:28.005745 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="918d3d63-22c7-433d-86c8-22747f6d3e77" path="/var/lib/kubelet/pods/918d3d63-22c7-433d-86c8-22747f6d3e77/volumes" Mar 13 09:36:28 crc kubenswrapper[4930]: I0313 09:36:28.006508 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62v7d" event={"ID":"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929","Type":"ContainerStarted","Data":"2deba84216225a676e1f124c0743a175b8a64979374dc1434ee67b9e91ef31ad"} Mar 13 09:36:28 crc kubenswrapper[4930]: I0313 09:36:28.007945 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:28 crc kubenswrapper[4930]: I0313 09:36:28.007965 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" event={"ID":"a2e994f9-023c-4cb7-b6b7-5a974a78e746","Type":"ContainerStarted","Data":"76e3a37a915bf1d9b57c8869108d42f44b5f2a831bfa3cef11772ed1074b344f"} Mar 13 09:36:28 crc kubenswrapper[4930]: I0313 09:36:28.007978 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" event={"ID":"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab","Type":"ContainerStarted","Data":"ca2507c79cb61939c1d8a9f126a31c5f6ee4a92ce38f71717c73143d90c3e90d"} Mar 13 09:36:28 crc kubenswrapper[4930]: I0313 09:36:28.033178 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-62v7d" podStartSLOduration=3.125551416 podStartE2EDuration="7.033155363s" podCreationTimestamp="2026-03-13 09:36:21 +0000 UTC" firstStartedPulling="2026-03-13 09:36:22.405647996 +0000 UTC m=+1423.155562673" lastFinishedPulling="2026-03-13 09:36:26.313251943 +0000 UTC m=+1427.063166620" observedRunningTime="2026-03-13 09:36:28.016065806 +0000 UTC m=+1428.765980473" watchObservedRunningTime="2026-03-13 09:36:28.033155363 +0000 UTC m=+1428.783070040" Mar 13 09:36:28 crc kubenswrapper[4930]: I0313 09:36:28.080801 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" podStartSLOduration=5.08078349 podStartE2EDuration="5.08078349s" podCreationTimestamp="2026-03-13 09:36:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:28.04870451 +0000 UTC m=+1428.798619187" watchObservedRunningTime="2026-03-13 09:36:28.08078349 +0000 UTC m=+1428.830698167" Mar 13 09:36:28 crc kubenswrapper[4930]: I0313 09:36:28.137582 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.010570 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" event={"ID":"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab","Type":"ContainerStarted","Data":"a17a89186b8ed6df4b2d47c457c9bb218b56a858fbc67b8f409eea56ef99481a"} Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.017573 4930 generic.go:334] "Generic (PLEG): container finished" podID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerID="07254621feced3ddd66802379857c0ce63823b857833b6250527f980d8badd16" exitCode=137 Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.017696 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7f00bfbf-551d-4730-bfeb-4c317e2a60e3","Type":"ContainerDied","Data":"07254621feced3ddd66802379857c0ce63823b857833b6250527f980d8badd16"} Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.232152 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.232824 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="178d7163-e4cc-4d25-9f42-2182928d43e4" containerName="glance-httpd" containerID="cri-o://c6936094e864674206cc13c48094456a401ee0af5803bd17c934cdc8e261e538" gracePeriod=30 Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.232707 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="178d7163-e4cc-4d25-9f42-2182928d43e4" containerName="glance-log" containerID="cri-o://b4d57afb3f9f4b4daa1177a0a7464e8a09b60d23e099df13c0deae7d6fbc4319" gracePeriod=30 Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.430136 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.465250 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-combined-ca-bundle\") pod \"aef74c6a-403d-4388-a873-43dca210f617\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.465328 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-ovndb-tls-certs\") pod \"aef74c6a-403d-4388-a873-43dca210f617\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.465513 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-httpd-config\") pod \"aef74c6a-403d-4388-a873-43dca210f617\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.466185 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvq4c\" (UniqueName: \"kubernetes.io/projected/aef74c6a-403d-4388-a873-43dca210f617-kube-api-access-tvq4c\") pod \"aef74c6a-403d-4388-a873-43dca210f617\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.466336 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-config\") pod \"aef74c6a-403d-4388-a873-43dca210f617\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.474702 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aef74c6a-403d-4388-a873-43dca210f617-kube-api-access-tvq4c" (OuterVolumeSpecName: "kube-api-access-tvq4c") pod "aef74c6a-403d-4388-a873-43dca210f617" (UID: "aef74c6a-403d-4388-a873-43dca210f617"). InnerVolumeSpecName "kube-api-access-tvq4c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.477558 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "aef74c6a-403d-4388-a873-43dca210f617" (UID: "aef74c6a-403d-4388-a873-43dca210f617"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.569331 4930 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.569371 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvq4c\" (UniqueName: \"kubernetes.io/projected/aef74c6a-403d-4388-a873-43dca210f617-kube-api-access-tvq4c\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.625713 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "aef74c6a-403d-4388-a873-43dca210f617" (UID: "aef74c6a-403d-4388-a873-43dca210f617"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.667148 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "aef74c6a-403d-4388-a873-43dca210f617" (UID: "aef74c6a-403d-4388-a873-43dca210f617"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.672857 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-config" (OuterVolumeSpecName: "config") pod "aef74c6a-403d-4388-a873-43dca210f617" (UID: "aef74c6a-403d-4388-a873-43dca210f617"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.674624 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-config\") pod \"aef74c6a-403d-4388-a873-43dca210f617\" (UID: \"aef74c6a-403d-4388-a873-43dca210f617\") " Mar 13 09:36:29 crc kubenswrapper[4930]: W0313 09:36:29.676341 4930 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/aef74c6a-403d-4388-a873-43dca210f617/volumes/kubernetes.io~secret/config Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.676389 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-config" (OuterVolumeSpecName: "config") pod "aef74c6a-403d-4388-a873-43dca210f617" (UID: "aef74c6a-403d-4388-a873-43dca210f617"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.676411 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.676450 4930 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:29 crc kubenswrapper[4930]: I0313 09:36:29.676462 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/aef74c6a-403d-4388-a873-43dca210f617-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.044624 4930 generic.go:334] "Generic (PLEG): container finished" podID="178d7163-e4cc-4d25-9f42-2182928d43e4" containerID="b4d57afb3f9f4b4daa1177a0a7464e8a09b60d23e099df13c0deae7d6fbc4319" exitCode=143 Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.044769 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"178d7163-e4cc-4d25-9f42-2182928d43e4","Type":"ContainerDied","Data":"b4d57afb3f9f4b4daa1177a0a7464e8a09b60d23e099df13c0deae7d6fbc4319"} Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.050590 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" event={"ID":"b0eff861-dcbe-4a46-97d1-bc1b4558a6ab","Type":"ContainerStarted","Data":"a53bd248f5b31849795a0b5994bc39c45148d43a374f08b6afe1c927fd03b1ec"} Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.051596 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.051625 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.058267 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6dcf4d8cb8-7d6r6" event={"ID":"aef74c6a-403d-4388-a873-43dca210f617","Type":"ContainerDied","Data":"daf06e22e6999efd5421ff9c6a42745bf71206ffd67ff1c23619fd861fb3d33a"} Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.058309 4930 scope.go:117] "RemoveContainer" containerID="35ba88787bfb2c888f48f45e752c3ba2753ebbdbab7fbe30723229f6d66e42fc" Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.058451 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6dcf4d8cb8-7d6r6" Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.121496 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6dcf4d8cb8-7d6r6"] Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.129807 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6dcf4d8cb8-7d6r6"] Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.134346 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" podStartSLOduration=4.134323045 podStartE2EDuration="4.134323045s" podCreationTimestamp="2026-03-13 09:36:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:30.109061004 +0000 UTC m=+1430.858975681" watchObservedRunningTime="2026-03-13 09:36:30.134323045 +0000 UTC m=+1430.884237722" Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.682554 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.683137 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="ceilometer-central-agent" containerID="cri-o://fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b" gracePeriod=30 Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.683200 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="proxy-httpd" containerID="cri-o://9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c" gracePeriod=30 Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.683273 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="ceilometer-notification-agent" containerID="cri-o://b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf" gracePeriod=30 Mar 13 09:36:30 crc kubenswrapper[4930]: I0313 09:36:30.683260 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="sg-core" containerID="cri-o://6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb" gracePeriod=30 Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.000890 4930 scope.go:117] "RemoveContainer" containerID="93155ac78aa8e7896ad1c995dc5aac22c196b3c9ecdcd86f67cbd7bfb454b789" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.085924 4930 generic.go:334] "Generic (PLEG): container finished" podID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerID="9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c" exitCode=0 Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.086247 4930 generic.go:334] "Generic (PLEG): container finished" podID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerID="6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb" exitCode=2 Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.086306 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1a220dc-49c4-4eb4-b05a-a07f3263b119","Type":"ContainerDied","Data":"9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c"} Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.086340 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1a220dc-49c4-4eb4-b05a-a07f3263b119","Type":"ContainerDied","Data":"6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb"} Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.091950 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"7f00bfbf-551d-4730-bfeb-4c317e2a60e3","Type":"ContainerDied","Data":"cfd90ac2096510a024278595d7cad533c17b686aa897e007a3cd76090289c2b7"} Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.092056 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfd90ac2096510a024278595d7cad533c17b686aa897e007a3cd76090289c2b7" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.162370 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.222228 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-config-data\") pod \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.222335 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-scripts\") pod \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.222968 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrtxs\" (UniqueName: \"kubernetes.io/projected/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-kube-api-access-zrtxs\") pod \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.223107 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-combined-ca-bundle\") pod \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.223188 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-logs\") pod \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.223256 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-etc-machine-id\") pod \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.223373 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-config-data-custom\") pod \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\" (UID: \"7f00bfbf-551d-4730-bfeb-4c317e2a60e3\") " Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.224585 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-logs" (OuterVolumeSpecName: "logs") pod "7f00bfbf-551d-4730-bfeb-4c317e2a60e3" (UID: "7f00bfbf-551d-4730-bfeb-4c317e2a60e3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.224759 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "7f00bfbf-551d-4730-bfeb-4c317e2a60e3" (UID: "7f00bfbf-551d-4730-bfeb-4c317e2a60e3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.240192 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-scripts" (OuterVolumeSpecName: "scripts") pod "7f00bfbf-551d-4730-bfeb-4c317e2a60e3" (UID: "7f00bfbf-551d-4730-bfeb-4c317e2a60e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.242645 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-kube-api-access-zrtxs" (OuterVolumeSpecName: "kube-api-access-zrtxs") pod "7f00bfbf-551d-4730-bfeb-4c317e2a60e3" (UID: "7f00bfbf-551d-4730-bfeb-4c317e2a60e3"). InnerVolumeSpecName "kube-api-access-zrtxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.245290 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7f00bfbf-551d-4730-bfeb-4c317e2a60e3" (UID: "7f00bfbf-551d-4730-bfeb-4c317e2a60e3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.288203 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7f00bfbf-551d-4730-bfeb-4c317e2a60e3" (UID: "7f00bfbf-551d-4730-bfeb-4c317e2a60e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.353672 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.354010 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrtxs\" (UniqueName: \"kubernetes.io/projected/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-kube-api-access-zrtxs\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.354029 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.354040 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.354052 4930 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.354063 4930 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.399102 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-config-data" (OuterVolumeSpecName: "config-data") pod "7f00bfbf-551d-4730-bfeb-4c317e2a60e3" (UID: "7f00bfbf-551d-4730-bfeb-4c317e2a60e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.455977 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7f00bfbf-551d-4730-bfeb-4c317e2a60e3-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.507529 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.508885 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:31 crc kubenswrapper[4930]: I0313 09:36:31.584482 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:31 crc kubenswrapper[4930]: E0313 09:36:31.775396 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1a220dc_49c4_4eb4_b05a_a07f3263b119.slice/crio-b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf.scope\": RecentStats: unable to find data in memory cache]" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.006730 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aef74c6a-403d-4388-a873-43dca210f617" path="/var/lib/kubelet/pods/aef74c6a-403d-4388-a873-43dca210f617/volumes" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.015387 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.073223 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-config-data\") pod \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.073587 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-scripts\") pod \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.073624 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1a220dc-49c4-4eb4-b05a-a07f3263b119-log-httpd\") pod \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.073783 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f72lc\" (UniqueName: \"kubernetes.io/projected/b1a220dc-49c4-4eb4-b05a-a07f3263b119-kube-api-access-f72lc\") pod \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.073896 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-combined-ca-bundle\") pod \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.073982 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1a220dc-49c4-4eb4-b05a-a07f3263b119-run-httpd\") pod \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.074014 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-sg-core-conf-yaml\") pod \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\" (UID: \"b1a220dc-49c4-4eb4-b05a-a07f3263b119\") " Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.076777 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1a220dc-49c4-4eb4-b05a-a07f3263b119-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b1a220dc-49c4-4eb4-b05a-a07f3263b119" (UID: "b1a220dc-49c4-4eb4-b05a-a07f3263b119"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.081298 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1a220dc-49c4-4eb4-b05a-a07f3263b119-kube-api-access-f72lc" (OuterVolumeSpecName: "kube-api-access-f72lc") pod "b1a220dc-49c4-4eb4-b05a-a07f3263b119" (UID: "b1a220dc-49c4-4eb4-b05a-a07f3263b119"). InnerVolumeSpecName "kube-api-access-f72lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.089168 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1a220dc-49c4-4eb4-b05a-a07f3263b119-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b1a220dc-49c4-4eb4-b05a-a07f3263b119" (UID: "b1a220dc-49c4-4eb4-b05a-a07f3263b119"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.091693 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-scripts" (OuterVolumeSpecName: "scripts") pod "b1a220dc-49c4-4eb4-b05a-a07f3263b119" (UID: "b1a220dc-49c4-4eb4-b05a-a07f3263b119"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.133961 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-558f5d58ff-jkrpd" event={"ID":"409ff6c8-9776-400c-9205-58658655a6a2","Type":"ContainerStarted","Data":"d797a82022c47a61191ce267b77101d6c6e4ce50ad8932bd81c8d2086452fbeb"} Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.134413 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.139469 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b1a220dc-49c4-4eb4-b05a-a07f3263b119" (UID: "b1a220dc-49c4-4eb4-b05a-a07f3263b119"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.180648 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f72lc\" (UniqueName: \"kubernetes.io/projected/b1a220dc-49c4-4eb4-b05a-a07f3263b119-kube-api-access-f72lc\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.180685 4930 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1a220dc-49c4-4eb4-b05a-a07f3263b119-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.180696 4930 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.180706 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.180715 4930 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b1a220dc-49c4-4eb4-b05a-a07f3263b119-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.186911 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" event={"ID":"7548398c-45a9-4165-81fe-3f15f68dc161","Type":"ContainerStarted","Data":"293f4ccae5256f85e7a553d009bf9d4a4ec9b9385f9b2557386cced5fe7d45ab"} Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.203572 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.210326 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-558f5d58ff-jkrpd" podStartSLOduration=3.046895167 podStartE2EDuration="8.210301897s" podCreationTimestamp="2026-03-13 09:36:24 +0000 UTC" firstStartedPulling="2026-03-13 09:36:25.837776846 +0000 UTC m=+1426.587691523" lastFinishedPulling="2026-03-13 09:36:31.001183566 +0000 UTC m=+1431.751098253" observedRunningTime="2026-03-13 09:36:32.159941708 +0000 UTC m=+1432.909856385" watchObservedRunningTime="2026-03-13 09:36:32.210301897 +0000 UTC m=+1432.960216574" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.267682 4930 generic.go:334] "Generic (PLEG): container finished" podID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerID="b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf" exitCode=0 Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.267713 4930 generic.go:334] "Generic (PLEG): container finished" podID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerID="fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b" exitCode=0 Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.268875 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.269358 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1a220dc-49c4-4eb4-b05a-a07f3263b119","Type":"ContainerDied","Data":"b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf"} Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.269384 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1a220dc-49c4-4eb4-b05a-a07f3263b119","Type":"ContainerDied","Data":"fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b"} Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.269395 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b1a220dc-49c4-4eb4-b05a-a07f3263b119","Type":"ContainerDied","Data":"5431d673d3d4cb2dd906e6abf20ec3740b0f4201cf87905cf5a8d64eb14630fa"} Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.269410 4930 scope.go:117] "RemoveContainer" containerID="9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.269492 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.280310 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" podStartSLOduration=3.6903943249999998 podStartE2EDuration="9.280285919s" podCreationTimestamp="2026-03-13 09:36:23 +0000 UTC" firstStartedPulling="2026-03-13 09:36:25.410599314 +0000 UTC m=+1426.160513991" lastFinishedPulling="2026-03-13 09:36:31.000490898 +0000 UTC m=+1431.750405585" observedRunningTime="2026-03-13 09:36:32.234794938 +0000 UTC m=+1432.984709615" watchObservedRunningTime="2026-03-13 09:36:32.280285919 +0000 UTC m=+1433.030200596" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.314024 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1a220dc-49c4-4eb4-b05a-a07f3263b119" (UID: "b1a220dc-49c4-4eb4-b05a-a07f3263b119"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.356263 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-config-data" (OuterVolumeSpecName: "config-data") pod "b1a220dc-49c4-4eb4-b05a-a07f3263b119" (UID: "b1a220dc-49c4-4eb4-b05a-a07f3263b119"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.371815 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.380202 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.385270 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.385294 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1a220dc-49c4-4eb4-b05a-a07f3263b119-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.477410 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="178d7163-e4cc-4d25-9f42-2182928d43e4" containerName="glance-httpd" probeResult="failure" output="Get \"https://10.217.0.198:9292/healthcheck\": dial tcp 10.217.0.198:9292: connect: connection refused" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.478039 4930 scope.go:117] "RemoveContainer" containerID="6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.481017 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/glance-default-internal-api-0" podUID="178d7163-e4cc-4d25-9f42-2182928d43e4" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.198:9292/healthcheck\": dial tcp 10.217.0.198:9292: connect: connection refused" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.486759 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.525875 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.552956 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 13 09:36:32 crc kubenswrapper[4930]: E0313 09:36:32.562575 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerName="cinder-api" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.562616 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerName="cinder-api" Mar 13 09:36:32 crc kubenswrapper[4930]: E0313 09:36:32.562661 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="ceilometer-central-agent" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.562669 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="ceilometer-central-agent" Mar 13 09:36:32 crc kubenswrapper[4930]: E0313 09:36:32.562687 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="proxy-httpd" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.562695 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="proxy-httpd" Mar 13 09:36:32 crc kubenswrapper[4930]: E0313 09:36:32.562726 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef74c6a-403d-4388-a873-43dca210f617" containerName="neutron-api" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.562736 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef74c6a-403d-4388-a873-43dca210f617" containerName="neutron-api" Mar 13 09:36:32 crc kubenswrapper[4930]: E0313 09:36:32.562769 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="sg-core" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.562777 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="sg-core" Mar 13 09:36:32 crc kubenswrapper[4930]: E0313 09:36:32.562818 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerName="cinder-api-log" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.562825 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerName="cinder-api-log" Mar 13 09:36:32 crc kubenswrapper[4930]: E0313 09:36:32.562866 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="ceilometer-notification-agent" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.562873 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="ceilometer-notification-agent" Mar 13 09:36:32 crc kubenswrapper[4930]: E0313 09:36:32.562889 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aef74c6a-403d-4388-a873-43dca210f617" containerName="neutron-httpd" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.562896 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="aef74c6a-403d-4388-a873-43dca210f617" containerName="neutron-httpd" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.575079 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerName="cinder-api-log" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.575153 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="ceilometer-notification-agent" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.575193 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="aef74c6a-403d-4388-a873-43dca210f617" containerName="neutron-httpd" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.575268 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="ceilometer-central-agent" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.575290 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="proxy-httpd" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.575311 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerName="cinder-api" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.575338 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="aef74c6a-403d-4388-a873-43dca210f617" containerName="neutron-api" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.575364 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" containerName="sg-core" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.595795 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.602294 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.602544 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.602809 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.604520 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.679084 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.692367 4930 scope.go:117] "RemoveContainer" containerID="b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.693172 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-scripts\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.693278 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d82ac246-17a4-4f81-acce-70aaa7d8f350-logs\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.693347 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.693387 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d82ac246-17a4-4f81-acce-70aaa7d8f350-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.693405 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.693546 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-config-data\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.693565 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-config-data-custom\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.693580 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f44x6\" (UniqueName: \"kubernetes.io/projected/d82ac246-17a4-4f81-acce-70aaa7d8f350-kube-api-access-f44x6\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.693666 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.696039 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.713864 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.717071 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.723116 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.723209 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.750719 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.764619 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-62v7d"] Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.784248 4930 scope.go:117] "RemoveContainer" containerID="fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.798624 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xmh5s\" (UniqueName: \"kubernetes.io/projected/e5091765-b1c8-43be-9579-d503faf84b6e-kube-api-access-xmh5s\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.798880 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.798925 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-config-data\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.798998 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e5091765-b1c8-43be-9579-d503faf84b6e-log-httpd\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.799053 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-scripts\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.799113 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.799160 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e5091765-b1c8-43be-9579-d503faf84b6e-run-httpd\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.799215 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d82ac246-17a4-4f81-acce-70aaa7d8f350-logs\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.799271 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.799320 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d82ac246-17a4-4f81-acce-70aaa7d8f350-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.799348 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.799384 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.799426 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-scripts\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.799504 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-config-data\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.799536 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-config-data-custom\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.799560 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f44x6\" (UniqueName: \"kubernetes.io/projected/d82ac246-17a4-4f81-acce-70aaa7d8f350-kube-api-access-f44x6\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.801602 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d82ac246-17a4-4f81-acce-70aaa7d8f350-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.805208 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d82ac246-17a4-4f81-acce-70aaa7d8f350-logs\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.823164 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-config-data-custom\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.823358 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-scripts\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.823804 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-public-tls-certs\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.823853 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.825813 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-config-data\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.835690 4930 scope.go:117] "RemoveContainer" containerID="9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.836155 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f44x6\" (UniqueName: \"kubernetes.io/projected/d82ac246-17a4-4f81-acce-70aaa7d8f350-kube-api-access-f44x6\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: E0313 09:36:32.840298 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c\": container with ID starting with 9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c not found: ID does not exist" containerID="9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.840367 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c"} err="failed to get container status \"9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c\": rpc error: code = NotFound desc = could not find container \"9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c\": container with ID starting with 9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c not found: ID does not exist" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.840400 4930 scope.go:117] "RemoveContainer" containerID="6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb" Mar 13 09:36:32 crc kubenswrapper[4930]: E0313 09:36:32.841020 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb\": container with ID starting with 6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb not found: ID does not exist" containerID="6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.841045 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb"} err="failed to get container status \"6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb\": rpc error: code = NotFound desc = could not find container \"6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb\": container with ID starting with 6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb not found: ID does not exist" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.841064 4930 scope.go:117] "RemoveContainer" containerID="b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.841227 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d82ac246-17a4-4f81-acce-70aaa7d8f350-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d82ac246-17a4-4f81-acce-70aaa7d8f350\") " pod="openstack/cinder-api-0" Mar 13 09:36:32 crc kubenswrapper[4930]: E0313 09:36:32.841650 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf\": container with ID starting with b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf not found: ID does not exist" containerID="b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.841676 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf"} err="failed to get container status \"b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf\": rpc error: code = NotFound desc = could not find container \"b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf\": container with ID starting with b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf not found: ID does not exist" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.841691 4930 scope.go:117] "RemoveContainer" containerID="fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b" Mar 13 09:36:32 crc kubenswrapper[4930]: E0313 09:36:32.841941 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b\": container with ID starting with fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b not found: ID does not exist" containerID="fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.841963 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b"} err="failed to get container status \"fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b\": rpc error: code = NotFound desc = could not find container \"fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b\": container with ID starting with fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b not found: ID does not exist" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.841978 4930 scope.go:117] "RemoveContainer" containerID="9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.842212 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c"} err="failed to get container status \"9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c\": rpc error: code = NotFound desc = could not find container \"9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c\": container with ID starting with 9b3b8d17390d29e236767076a3c14e698989fcdfa9e2134761e665638912324c not found: ID does not exist" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.842229 4930 scope.go:117] "RemoveContainer" containerID="6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.843363 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb"} err="failed to get container status \"6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb\": rpc error: code = NotFound desc = could not find container \"6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb\": container with ID starting with 6a44f034e267fd8e7022086fc1d8539d526180780f6f9554bd63028cca949ebb not found: ID does not exist" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.843386 4930 scope.go:117] "RemoveContainer" containerID="b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.844012 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf"} err="failed to get container status \"b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf\": rpc error: code = NotFound desc = could not find container \"b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf\": container with ID starting with b6fe0baffee90de2d16ff9109402a1de0f7aded8f94c04c2798fc80116671ebf not found: ID does not exist" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.844030 4930 scope.go:117] "RemoveContainer" containerID="fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.844250 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b"} err="failed to get container status \"fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b\": rpc error: code = NotFound desc = could not find container \"fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b\": container with ID starting with fc293f635f09644b1e07151e96ff319a5c838a006ec484a906efda5b79186f7b not found: ID does not exist" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.901859 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e5091765-b1c8-43be-9579-d503faf84b6e-log-httpd\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.902021 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.902426 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e5091765-b1c8-43be-9579-d503faf84b6e-run-httpd\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.902638 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.902713 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-scripts\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.902893 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xmh5s\" (UniqueName: \"kubernetes.io/projected/e5091765-b1c8-43be-9579-d503faf84b6e-kube-api-access-xmh5s\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.903042 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-config-data\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.907089 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e5091765-b1c8-43be-9579-d503faf84b6e-run-httpd\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.907378 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e5091765-b1c8-43be-9579-d503faf84b6e-log-httpd\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.909003 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-config-data\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.912791 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.916196 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.919059 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-scripts\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.929256 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xmh5s\" (UniqueName: \"kubernetes.io/projected/e5091765-b1c8-43be-9579-d503faf84b6e-kube-api-access-xmh5s\") pod \"ceilometer-0\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " pod="openstack/ceilometer-0" Mar 13 09:36:32 crc kubenswrapper[4930]: I0313 09:36:32.934331 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.060601 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.122549 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-c6b696d47-c76xv"] Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.131266 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.143707 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-c6b696d47-c76xv"] Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.157773 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-66d7f6c554-7d4pc"] Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.159672 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.169469 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-66d7f6c554-7d4pc"] Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.180770 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7889c489fb-hj2f9"] Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.184996 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.209051 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-combined-ca-bundle\") pod \"heat-engine-c6b696d47-c76xv\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.209142 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-config-data\") pod \"heat-api-7889c489fb-hj2f9\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.209215 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-config-data\") pod \"heat-engine-c6b696d47-c76xv\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.209326 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-combined-ca-bundle\") pod \"heat-api-7889c489fb-hj2f9\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.209370 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-config-data-custom\") pod \"heat-engine-c6b696d47-c76xv\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.209408 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-config-data-custom\") pod \"heat-cfnapi-66d7f6c554-7d4pc\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.209483 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9n6vz\" (UniqueName: \"kubernetes.io/projected/17c2545f-f1f1-4c0a-a689-73202f48ddf6-kube-api-access-9n6vz\") pod \"heat-cfnapi-66d7f6c554-7d4pc\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.209557 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-combined-ca-bundle\") pod \"heat-cfnapi-66d7f6c554-7d4pc\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.209597 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6p8g8\" (UniqueName: \"kubernetes.io/projected/f0868712-845b-4803-b7a7-44aaa23ad284-kube-api-access-6p8g8\") pod \"heat-engine-c6b696d47-c76xv\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.209619 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-config-data-custom\") pod \"heat-api-7889c489fb-hj2f9\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.209734 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdhkg\" (UniqueName: \"kubernetes.io/projected/304ba316-73ff-420c-af72-945f29aa5e2e-kube-api-access-zdhkg\") pod \"heat-api-7889c489fb-hj2f9\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.209891 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-config-data\") pod \"heat-cfnapi-66d7f6c554-7d4pc\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.215234 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7889c489fb-hj2f9"] Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.286867 4930 generic.go:334] "Generic (PLEG): container finished" podID="178d7163-e4cc-4d25-9f42-2182928d43e4" containerID="c6936094e864674206cc13c48094456a401ee0af5803bd17c934cdc8e261e538" exitCode=0 Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.287813 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"178d7163-e4cc-4d25-9f42-2182928d43e4","Type":"ContainerDied","Data":"c6936094e864674206cc13c48094456a401ee0af5803bd17c934cdc8e261e538"} Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.311656 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-config-data\") pod \"heat-cfnapi-66d7f6c554-7d4pc\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.311702 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-combined-ca-bundle\") pod \"heat-engine-c6b696d47-c76xv\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.311739 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-config-data\") pod \"heat-api-7889c489fb-hj2f9\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.311822 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-config-data\") pod \"heat-engine-c6b696d47-c76xv\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.311858 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-combined-ca-bundle\") pod \"heat-api-7889c489fb-hj2f9\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.311886 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-config-data-custom\") pod \"heat-engine-c6b696d47-c76xv\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.311918 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-config-data-custom\") pod \"heat-cfnapi-66d7f6c554-7d4pc\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.311938 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9n6vz\" (UniqueName: \"kubernetes.io/projected/17c2545f-f1f1-4c0a-a689-73202f48ddf6-kube-api-access-9n6vz\") pod \"heat-cfnapi-66d7f6c554-7d4pc\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.312019 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-combined-ca-bundle\") pod \"heat-cfnapi-66d7f6c554-7d4pc\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.312051 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6p8g8\" (UniqueName: \"kubernetes.io/projected/f0868712-845b-4803-b7a7-44aaa23ad284-kube-api-access-6p8g8\") pod \"heat-engine-c6b696d47-c76xv\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.312069 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-config-data-custom\") pod \"heat-api-7889c489fb-hj2f9\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.312223 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdhkg\" (UniqueName: \"kubernetes.io/projected/304ba316-73ff-420c-af72-945f29aa5e2e-kube-api-access-zdhkg\") pod \"heat-api-7889c489fb-hj2f9\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.318497 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-config-data\") pod \"heat-api-7889c489fb-hj2f9\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.319063 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-config-data-custom\") pod \"heat-cfnapi-66d7f6c554-7d4pc\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.324400 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-combined-ca-bundle\") pod \"heat-engine-c6b696d47-c76xv\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.326003 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-combined-ca-bundle\") pod \"heat-api-7889c489fb-hj2f9\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.327354 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-config-data\") pod \"heat-cfnapi-66d7f6c554-7d4pc\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.328237 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-config-data-custom\") pod \"heat-api-7889c489fb-hj2f9\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.331404 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-config-data-custom\") pod \"heat-engine-c6b696d47-c76xv\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.333976 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-combined-ca-bundle\") pod \"heat-cfnapi-66d7f6c554-7d4pc\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.338354 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6p8g8\" (UniqueName: \"kubernetes.io/projected/f0868712-845b-4803-b7a7-44aaa23ad284-kube-api-access-6p8g8\") pod \"heat-engine-c6b696d47-c76xv\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.339016 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdhkg\" (UniqueName: \"kubernetes.io/projected/304ba316-73ff-420c-af72-945f29aa5e2e-kube-api-access-zdhkg\") pod \"heat-api-7889c489fb-hj2f9\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.340382 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9n6vz\" (UniqueName: \"kubernetes.io/projected/17c2545f-f1f1-4c0a-a689-73202f48ddf6-kube-api-access-9n6vz\") pod \"heat-cfnapi-66d7f6c554-7d4pc\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.343754 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-config-data\") pod \"heat-engine-c6b696d47-c76xv\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.370273 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/cinder-api-0" podUID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" containerName="cinder-api" probeResult="failure" output="Get \"http://10.217.0.215:8776/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.488531 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.507034 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.517744 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.988711 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f00bfbf-551d-4730-bfeb-4c317e2a60e3" path="/var/lib/kubelet/pods/7f00bfbf-551d-4730-bfeb-4c317e2a60e3/volumes" Mar 13 09:36:33 crc kubenswrapper[4930]: I0313 09:36:33.989660 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1a220dc-49c4-4eb4-b05a-a07f3263b119" path="/var/lib/kubelet/pods/b1a220dc-49c4-4eb4-b05a-a07f3263b119/volumes" Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.298680 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-62v7d" podUID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" containerName="registry-server" containerID="cri-o://2deba84216225a676e1f124c0743a175b8a64979374dc1434ee67b9e91ef31ad" gracePeriod=2 Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.390582 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.468340 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7wx75"] Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.468710 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" podUID="c765eaa6-eb16-4c09-b845-430cfd29e6d6" containerName="dnsmasq-dns" containerID="cri-o://ff2905cd51a663f0a9d93f7c4daaf6804a2402ddf3b658878cb899db630b23cd" gracePeriod=10 Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.855129 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-f4dfb5f5b-pw8zp"] Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.855387 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" podUID="7548398c-45a9-4165-81fe-3f15f68dc161" containerName="heat-cfnapi" containerID="cri-o://293f4ccae5256f85e7a553d009bf9d4a4ec9b9385f9b2557386cced5fe7d45ab" gracePeriod=60 Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.881520 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-558f5d58ff-jkrpd"] Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.881776 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-558f5d58ff-jkrpd" podUID="409ff6c8-9776-400c-9205-58658655a6a2" containerName="heat-api" containerID="cri-o://d797a82022c47a61191ce267b77101d6c6e4ce50ad8932bd81c8d2086452fbeb" gracePeriod=60 Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.932603 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-86b9fc57f7-vjhx9"] Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.934410 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.937298 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.937449 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.958650 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-77757997f5-wlmc4"] Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.960159 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.964384 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Mar 13 09:36:34 crc kubenswrapper[4930]: I0313 09:36:34.964706 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.005660 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-86b9fc57f7-vjhx9"] Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.042501 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-77757997f5-wlmc4"] Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.075273 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-config-data-custom\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.075339 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-config-data\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.075393 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-combined-ca-bundle\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.075457 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-internal-tls-certs\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.075480 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-combined-ca-bundle\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.075600 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkbp2\" (UniqueName: \"kubernetes.io/projected/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-kube-api-access-mkbp2\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.075712 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-config-data\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.075881 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-internal-tls-certs\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.076011 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hz4tm\" (UniqueName: \"kubernetes.io/projected/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-kube-api-access-hz4tm\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.076057 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-public-tls-certs\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.076145 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-config-data-custom\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.076189 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-public-tls-certs\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.177747 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-config-data-custom\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.177796 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-config-data\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.177834 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-combined-ca-bundle\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.177861 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-combined-ca-bundle\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.177876 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-internal-tls-certs\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.177897 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkbp2\" (UniqueName: \"kubernetes.io/projected/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-kube-api-access-mkbp2\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.177925 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-config-data\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.177980 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-internal-tls-certs\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.178020 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hz4tm\" (UniqueName: \"kubernetes.io/projected/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-kube-api-access-hz4tm\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.178038 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-public-tls-certs\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.178066 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-config-data-custom\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.178081 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-public-tls-certs\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.186560 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-config-data-custom\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.195226 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-internal-tls-certs\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.195421 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-public-tls-certs\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.197473 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-config-data\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.201426 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-config-data-custom\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.205496 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-config-data\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.206090 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-combined-ca-bundle\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.210219 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-public-tls-certs\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.219438 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-internal-tls-certs\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.220113 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-combined-ca-bundle\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.234557 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkbp2\" (UniqueName: \"kubernetes.io/projected/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-kube-api-access-mkbp2\") pod \"heat-api-77757997f5-wlmc4\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.235237 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hz4tm\" (UniqueName: \"kubernetes.io/projected/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-kube-api-access-hz4tm\") pod \"heat-cfnapi-86b9fc57f7-vjhx9\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.267007 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.321823 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.336335 4930 generic.go:334] "Generic (PLEG): container finished" podID="c765eaa6-eb16-4c09-b845-430cfd29e6d6" containerID="ff2905cd51a663f0a9d93f7c4daaf6804a2402ddf3b658878cb899db630b23cd" exitCode=0 Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.336397 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" event={"ID":"c765eaa6-eb16-4c09-b845-430cfd29e6d6","Type":"ContainerDied","Data":"ff2905cd51a663f0a9d93f7c4daaf6804a2402ddf3b658878cb899db630b23cd"} Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.343832 4930 generic.go:334] "Generic (PLEG): container finished" podID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" containerID="2deba84216225a676e1f124c0743a175b8a64979374dc1434ee67b9e91ef31ad" exitCode=0 Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.344111 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62v7d" event={"ID":"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929","Type":"ContainerDied","Data":"2deba84216225a676e1f124c0743a175b8a64979374dc1434ee67b9e91ef31ad"} Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.529901 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.530386 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d95152a3-5e16-464b-9785-88e629cc31f9" containerName="glance-log" containerID="cri-o://995636287c1ceecdc41a41988d0a3367c860c034253f4dffe261994bdfc30a8f" gracePeriod=30 Mar 13 09:36:35 crc kubenswrapper[4930]: I0313 09:36:35.530534 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="d95152a3-5e16-464b-9785-88e629cc31f9" containerName="glance-httpd" containerID="cri-o://29fb594a8475f74faf9b69578f86c18f499f27eb837df8c5196f92b52408d168" gracePeriod=30 Mar 13 09:36:36 crc kubenswrapper[4930]: I0313 09:36:36.358721 4930 generic.go:334] "Generic (PLEG): container finished" podID="d95152a3-5e16-464b-9785-88e629cc31f9" containerID="995636287c1ceecdc41a41988d0a3367c860c034253f4dffe261994bdfc30a8f" exitCode=143 Mar 13 09:36:36 crc kubenswrapper[4930]: I0313 09:36:36.358784 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d95152a3-5e16-464b-9785-88e629cc31f9","Type":"ContainerDied","Data":"995636287c1ceecdc41a41988d0a3367c860c034253f4dffe261994bdfc30a8f"} Mar 13 09:36:36 crc kubenswrapper[4930]: I0313 09:36:36.360789 4930 generic.go:334] "Generic (PLEG): container finished" podID="409ff6c8-9776-400c-9205-58658655a6a2" containerID="d797a82022c47a61191ce267b77101d6c6e4ce50ad8932bd81c8d2086452fbeb" exitCode=0 Mar 13 09:36:36 crc kubenswrapper[4930]: I0313 09:36:36.360837 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-558f5d58ff-jkrpd" event={"ID":"409ff6c8-9776-400c-9205-58658655a6a2","Type":"ContainerDied","Data":"d797a82022c47a61191ce267b77101d6c6e4ce50ad8932bd81c8d2086452fbeb"} Mar 13 09:36:36 crc kubenswrapper[4930]: I0313 09:36:36.364799 4930 generic.go:334] "Generic (PLEG): container finished" podID="7548398c-45a9-4165-81fe-3f15f68dc161" containerID="293f4ccae5256f85e7a553d009bf9d4a4ec9b9385f9b2557386cced5fe7d45ab" exitCode=0 Mar 13 09:36:36 crc kubenswrapper[4930]: I0313 09:36:36.364845 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" event={"ID":"7548398c-45a9-4165-81fe-3f15f68dc161","Type":"ContainerDied","Data":"293f4ccae5256f85e7a553d009bf9d4a4ec9b9385f9b2557386cced5fe7d45ab"} Mar 13 09:36:36 crc kubenswrapper[4930]: I0313 09:36:36.532791 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-bd7b6bc5f-hcdp4" Mar 13 09:36:37 crc kubenswrapper[4930]: I0313 09:36:37.307817 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bsl74" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerName="registry-server" probeResult="failure" output=< Mar 13 09:36:37 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:36:37 crc kubenswrapper[4930]: > Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.224667 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" podUID="c765eaa6-eb16-4c09-b845-430cfd29e6d6" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.214:5353: connect: connection refused" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.649262 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-lzrnq"] Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.651112 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lzrnq" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.670138 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lzrnq"] Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.765489 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-b5f3-account-create-update-9zsnn"] Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.767290 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b5f3-account-create-update-9zsnn" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.771379 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.779657 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gq2ll\" (UniqueName: \"kubernetes.io/projected/9e12cb85-3947-4c60-96a6-8ac07f5e63e8-kube-api-access-gq2ll\") pod \"nova-api-db-create-lzrnq\" (UID: \"9e12cb85-3947-4c60-96a6-8ac07f5e63e8\") " pod="openstack/nova-api-db-create-lzrnq" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.779771 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e12cb85-3947-4c60-96a6-8ac07f5e63e8-operator-scripts\") pod \"nova-api-db-create-lzrnq\" (UID: \"9e12cb85-3947-4c60-96a6-8ac07f5e63e8\") " pod="openstack/nova-api-db-create-lzrnq" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.783807 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-qbssl"] Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.785609 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qbssl" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.800324 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b5f3-account-create-update-9zsnn"] Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.814492 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-qbssl"] Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.881427 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e12cb85-3947-4c60-96a6-8ac07f5e63e8-operator-scripts\") pod \"nova-api-db-create-lzrnq\" (UID: \"9e12cb85-3947-4c60-96a6-8ac07f5e63e8\") " pod="openstack/nova-api-db-create-lzrnq" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.881557 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grt5l\" (UniqueName: \"kubernetes.io/projected/ec86dec2-eb5a-497e-bba2-d1b96d80c65c-kube-api-access-grt5l\") pod \"nova-cell0-db-create-qbssl\" (UID: \"ec86dec2-eb5a-497e-bba2-d1b96d80c65c\") " pod="openstack/nova-cell0-db-create-qbssl" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.881629 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gq2ll\" (UniqueName: \"kubernetes.io/projected/9e12cb85-3947-4c60-96a6-8ac07f5e63e8-kube-api-access-gq2ll\") pod \"nova-api-db-create-lzrnq\" (UID: \"9e12cb85-3947-4c60-96a6-8ac07f5e63e8\") " pod="openstack/nova-api-db-create-lzrnq" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.881658 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a92113b-21a5-4db3-9708-589fc0aa57e8-operator-scripts\") pod \"nova-api-b5f3-account-create-update-9zsnn\" (UID: \"4a92113b-21a5-4db3-9708-589fc0aa57e8\") " pod="openstack/nova-api-b5f3-account-create-update-9zsnn" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.881714 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h84lg\" (UniqueName: \"kubernetes.io/projected/4a92113b-21a5-4db3-9708-589fc0aa57e8-kube-api-access-h84lg\") pod \"nova-api-b5f3-account-create-update-9zsnn\" (UID: \"4a92113b-21a5-4db3-9708-589fc0aa57e8\") " pod="openstack/nova-api-b5f3-account-create-update-9zsnn" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.881740 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec86dec2-eb5a-497e-bba2-d1b96d80c65c-operator-scripts\") pod \"nova-cell0-db-create-qbssl\" (UID: \"ec86dec2-eb5a-497e-bba2-d1b96d80c65c\") " pod="openstack/nova-cell0-db-create-qbssl" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.882604 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e12cb85-3947-4c60-96a6-8ac07f5e63e8-operator-scripts\") pod \"nova-api-db-create-lzrnq\" (UID: \"9e12cb85-3947-4c60-96a6-8ac07f5e63e8\") " pod="openstack/nova-api-db-create-lzrnq" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.906480 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gq2ll\" (UniqueName: \"kubernetes.io/projected/9e12cb85-3947-4c60-96a6-8ac07f5e63e8-kube-api-access-gq2ll\") pod \"nova-api-db-create-lzrnq\" (UID: \"9e12cb85-3947-4c60-96a6-8ac07f5e63e8\") " pod="openstack/nova-api-db-create-lzrnq" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.970569 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-xxmsm"] Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.972716 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xxmsm" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.983798 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grt5l\" (UniqueName: \"kubernetes.io/projected/ec86dec2-eb5a-497e-bba2-d1b96d80c65c-kube-api-access-grt5l\") pod \"nova-cell0-db-create-qbssl\" (UID: \"ec86dec2-eb5a-497e-bba2-d1b96d80c65c\") " pod="openstack/nova-cell0-db-create-qbssl" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.983897 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a92113b-21a5-4db3-9708-589fc0aa57e8-operator-scripts\") pod \"nova-api-b5f3-account-create-update-9zsnn\" (UID: \"4a92113b-21a5-4db3-9708-589fc0aa57e8\") " pod="openstack/nova-api-b5f3-account-create-update-9zsnn" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.983953 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h84lg\" (UniqueName: \"kubernetes.io/projected/4a92113b-21a5-4db3-9708-589fc0aa57e8-kube-api-access-h84lg\") pod \"nova-api-b5f3-account-create-update-9zsnn\" (UID: \"4a92113b-21a5-4db3-9708-589fc0aa57e8\") " pod="openstack/nova-api-b5f3-account-create-update-9zsnn" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.983978 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec86dec2-eb5a-497e-bba2-d1b96d80c65c-operator-scripts\") pod \"nova-cell0-db-create-qbssl\" (UID: \"ec86dec2-eb5a-497e-bba2-d1b96d80c65c\") " pod="openstack/nova-cell0-db-create-qbssl" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.984661 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec86dec2-eb5a-497e-bba2-d1b96d80c65c-operator-scripts\") pod \"nova-cell0-db-create-qbssl\" (UID: \"ec86dec2-eb5a-497e-bba2-d1b96d80c65c\") " pod="openstack/nova-cell0-db-create-qbssl" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.985389 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a92113b-21a5-4db3-9708-589fc0aa57e8-operator-scripts\") pod \"nova-api-b5f3-account-create-update-9zsnn\" (UID: \"4a92113b-21a5-4db3-9708-589fc0aa57e8\") " pod="openstack/nova-api-b5f3-account-create-update-9zsnn" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.985575 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-7e6b-account-create-update-tgmkk"] Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.987118 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.990956 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lzrnq" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.992731 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 13 09:36:38 crc kubenswrapper[4930]: I0313 09:36:38.994465 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7e6b-account-create-update-tgmkk"] Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.009022 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xxmsm"] Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.034772 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h84lg\" (UniqueName: \"kubernetes.io/projected/4a92113b-21a5-4db3-9708-589fc0aa57e8-kube-api-access-h84lg\") pod \"nova-api-b5f3-account-create-update-9zsnn\" (UID: \"4a92113b-21a5-4db3-9708-589fc0aa57e8\") " pod="openstack/nova-api-b5f3-account-create-update-9zsnn" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.043988 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grt5l\" (UniqueName: \"kubernetes.io/projected/ec86dec2-eb5a-497e-bba2-d1b96d80c65c-kube-api-access-grt5l\") pod \"nova-cell0-db-create-qbssl\" (UID: \"ec86dec2-eb5a-497e-bba2-d1b96d80c65c\") " pod="openstack/nova-cell0-db-create-qbssl" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.086122 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a15fab1-c914-4555-813a-ab9552b29a22-operator-scripts\") pod \"nova-cell0-7e6b-account-create-update-tgmkk\" (UID: \"4a15fab1-c914-4555-813a-ab9552b29a22\") " pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.086231 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4ml6\" (UniqueName: \"kubernetes.io/projected/b3c2e237-7f4c-48a2-8a8d-f5838192decc-kube-api-access-f4ml6\") pod \"nova-cell1-db-create-xxmsm\" (UID: \"b3c2e237-7f4c-48a2-8a8d-f5838192decc\") " pod="openstack/nova-cell1-db-create-xxmsm" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.086326 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqksg\" (UniqueName: \"kubernetes.io/projected/4a15fab1-c914-4555-813a-ab9552b29a22-kube-api-access-dqksg\") pod \"nova-cell0-7e6b-account-create-update-tgmkk\" (UID: \"4a15fab1-c914-4555-813a-ab9552b29a22\") " pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.086419 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3c2e237-7f4c-48a2-8a8d-f5838192decc-operator-scripts\") pod \"nova-cell1-db-create-xxmsm\" (UID: \"b3c2e237-7f4c-48a2-8a8d-f5838192decc\") " pod="openstack/nova-cell1-db-create-xxmsm" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.105708 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b5f3-account-create-update-9zsnn" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.130726 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qbssl" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.189462 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a15fab1-c914-4555-813a-ab9552b29a22-operator-scripts\") pod \"nova-cell0-7e6b-account-create-update-tgmkk\" (UID: \"4a15fab1-c914-4555-813a-ab9552b29a22\") " pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.189586 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4ml6\" (UniqueName: \"kubernetes.io/projected/b3c2e237-7f4c-48a2-8a8d-f5838192decc-kube-api-access-f4ml6\") pod \"nova-cell1-db-create-xxmsm\" (UID: \"b3c2e237-7f4c-48a2-8a8d-f5838192decc\") " pod="openstack/nova-cell1-db-create-xxmsm" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.189669 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqksg\" (UniqueName: \"kubernetes.io/projected/4a15fab1-c914-4555-813a-ab9552b29a22-kube-api-access-dqksg\") pod \"nova-cell0-7e6b-account-create-update-tgmkk\" (UID: \"4a15fab1-c914-4555-813a-ab9552b29a22\") " pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.189779 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3c2e237-7f4c-48a2-8a8d-f5838192decc-operator-scripts\") pod \"nova-cell1-db-create-xxmsm\" (UID: \"b3c2e237-7f4c-48a2-8a8d-f5838192decc\") " pod="openstack/nova-cell1-db-create-xxmsm" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.190418 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3c2e237-7f4c-48a2-8a8d-f5838192decc-operator-scripts\") pod \"nova-cell1-db-create-xxmsm\" (UID: \"b3c2e237-7f4c-48a2-8a8d-f5838192decc\") " pod="openstack/nova-cell1-db-create-xxmsm" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.191901 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-122f-account-create-update-9988g"] Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.196132 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-122f-account-create-update-9988g" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.202647 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.203934 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a15fab1-c914-4555-813a-ab9552b29a22-operator-scripts\") pod \"nova-cell0-7e6b-account-create-update-tgmkk\" (UID: \"4a15fab1-c914-4555-813a-ab9552b29a22\") " pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.204778 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-122f-account-create-update-9988g"] Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.206907 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4ml6\" (UniqueName: \"kubernetes.io/projected/b3c2e237-7f4c-48a2-8a8d-f5838192decc-kube-api-access-f4ml6\") pod \"nova-cell1-db-create-xxmsm\" (UID: \"b3c2e237-7f4c-48a2-8a8d-f5838192decc\") " pod="openstack/nova-cell1-db-create-xxmsm" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.210239 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqksg\" (UniqueName: \"kubernetes.io/projected/4a15fab1-c914-4555-813a-ab9552b29a22-kube-api-access-dqksg\") pod \"nova-cell0-7e6b-account-create-update-tgmkk\" (UID: \"4a15fab1-c914-4555-813a-ab9552b29a22\") " pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.292321 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5942b4a-ef6d-40ca-b67f-e7965513004d-operator-scripts\") pod \"nova-cell1-122f-account-create-update-9988g\" (UID: \"f5942b4a-ef6d-40ca-b67f-e7965513004d\") " pod="openstack/nova-cell1-122f-account-create-update-9988g" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.292418 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rsdt8\" (UniqueName: \"kubernetes.io/projected/f5942b4a-ef6d-40ca-b67f-e7965513004d-kube-api-access-rsdt8\") pod \"nova-cell1-122f-account-create-update-9988g\" (UID: \"f5942b4a-ef6d-40ca-b67f-e7965513004d\") " pod="openstack/nova-cell1-122f-account-create-update-9988g" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.393183 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" podUID="7548398c-45a9-4165-81fe-3f15f68dc161" containerName="heat-cfnapi" probeResult="failure" output="Get \"http://10.217.0.224:8000/healthcheck\": dial tcp 10.217.0.224:8000: connect: connection refused" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.395882 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5942b4a-ef6d-40ca-b67f-e7965513004d-operator-scripts\") pod \"nova-cell1-122f-account-create-update-9988g\" (UID: \"f5942b4a-ef6d-40ca-b67f-e7965513004d\") " pod="openstack/nova-cell1-122f-account-create-update-9988g" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.395972 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rsdt8\" (UniqueName: \"kubernetes.io/projected/f5942b4a-ef6d-40ca-b67f-e7965513004d-kube-api-access-rsdt8\") pod \"nova-cell1-122f-account-create-update-9988g\" (UID: \"f5942b4a-ef6d-40ca-b67f-e7965513004d\") " pod="openstack/nova-cell1-122f-account-create-update-9988g" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.396825 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5942b4a-ef6d-40ca-b67f-e7965513004d-operator-scripts\") pod \"nova-cell1-122f-account-create-update-9988g\" (UID: \"f5942b4a-ef6d-40ca-b67f-e7965513004d\") " pod="openstack/nova-cell1-122f-account-create-update-9988g" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.420265 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rsdt8\" (UniqueName: \"kubernetes.io/projected/f5942b4a-ef6d-40ca-b67f-e7965513004d-kube-api-access-rsdt8\") pod \"nova-cell1-122f-account-create-update-9988g\" (UID: \"f5942b4a-ef6d-40ca-b67f-e7965513004d\") " pod="openstack/nova-cell1-122f-account-create-update-9988g" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.420996 4930 generic.go:334] "Generic (PLEG): container finished" podID="d95152a3-5e16-464b-9785-88e629cc31f9" containerID="29fb594a8475f74faf9b69578f86c18f499f27eb837df8c5196f92b52408d168" exitCode=0 Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.421039 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d95152a3-5e16-464b-9785-88e629cc31f9","Type":"ContainerDied","Data":"29fb594a8475f74faf9b69578f86c18f499f27eb837df8c5196f92b52408d168"} Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.458881 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xxmsm" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.472367 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.669752 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-122f-account-create-update-9988g" Mar 13 09:36:39 crc kubenswrapper[4930]: I0313 09:36:39.755338 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-558f5d58ff-jkrpd" podUID="409ff6c8-9776-400c-9205-58658655a6a2" containerName="heat-api" probeResult="failure" output="Get \"http://10.217.0.226:8004/healthcheck\": dial tcp 10.217.0.226:8004: connect: connection refused" Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.126954 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.751955 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.837419 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2srw\" (UniqueName: \"kubernetes.io/projected/178d7163-e4cc-4d25-9f42-2182928d43e4-kube-api-access-g2srw\") pod \"178d7163-e4cc-4d25-9f42-2182928d43e4\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.837527 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/178d7163-e4cc-4d25-9f42-2182928d43e4-httpd-run\") pod \"178d7163-e4cc-4d25-9f42-2182928d43e4\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.837632 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-scripts\") pod \"178d7163-e4cc-4d25-9f42-2182928d43e4\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.837683 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-config-data\") pod \"178d7163-e4cc-4d25-9f42-2182928d43e4\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.837717 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-internal-tls-certs\") pod \"178d7163-e4cc-4d25-9f42-2182928d43e4\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.838205 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"178d7163-e4cc-4d25-9f42-2182928d43e4\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.838251 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/178d7163-e4cc-4d25-9f42-2182928d43e4-logs\") pod \"178d7163-e4cc-4d25-9f42-2182928d43e4\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.838357 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-combined-ca-bundle\") pod \"178d7163-e4cc-4d25-9f42-2182928d43e4\" (UID: \"178d7163-e4cc-4d25-9f42-2182928d43e4\") " Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.838201 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/178d7163-e4cc-4d25-9f42-2182928d43e4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "178d7163-e4cc-4d25-9f42-2182928d43e4" (UID: "178d7163-e4cc-4d25-9f42-2182928d43e4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.838589 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/178d7163-e4cc-4d25-9f42-2182928d43e4-logs" (OuterVolumeSpecName: "logs") pod "178d7163-e4cc-4d25-9f42-2182928d43e4" (UID: "178d7163-e4cc-4d25-9f42-2182928d43e4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.839424 4930 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/178d7163-e4cc-4d25-9f42-2182928d43e4-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.839469 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/178d7163-e4cc-4d25-9f42-2182928d43e4-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.849796 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/178d7163-e4cc-4d25-9f42-2182928d43e4-kube-api-access-g2srw" (OuterVolumeSpecName: "kube-api-access-g2srw") pod "178d7163-e4cc-4d25-9f42-2182928d43e4" (UID: "178d7163-e4cc-4d25-9f42-2182928d43e4"). InnerVolumeSpecName "kube-api-access-g2srw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.851349 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-scripts" (OuterVolumeSpecName: "scripts") pod "178d7163-e4cc-4d25-9f42-2182928d43e4" (UID: "178d7163-e4cc-4d25-9f42-2182928d43e4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.905589 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f" (OuterVolumeSpecName: "glance") pod "178d7163-e4cc-4d25-9f42-2182928d43e4" (UID: "178d7163-e4cc-4d25-9f42-2182928d43e4"). InnerVolumeSpecName "pvc-893414a0-cbec-42bd-a89f-5c787585718f". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.943057 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2srw\" (UniqueName: \"kubernetes.io/projected/178d7163-e4cc-4d25-9f42-2182928d43e4-kube-api-access-g2srw\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.943089 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.943118 4930 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") on node \"crc\" " Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.981676 4930 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 09:36:40 crc kubenswrapper[4930]: I0313 09:36:40.982159 4930 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-893414a0-cbec-42bd-a89f-5c787585718f" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f") on node "crc" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.020103 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "178d7163-e4cc-4d25-9f42-2182928d43e4" (UID: "178d7163-e4cc-4d25-9f42-2182928d43e4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.034574 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-config-data" (OuterVolumeSpecName: "config-data") pod "178d7163-e4cc-4d25-9f42-2182928d43e4" (UID: "178d7163-e4cc-4d25-9f42-2182928d43e4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.046658 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.046685 4930 reconciler_common.go:293] "Volume detached for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.046699 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.063506 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "178d7163-e4cc-4d25-9f42-2182928d43e4" (UID: "178d7163-e4cc-4d25-9f42-2182928d43e4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.159439 4930 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/178d7163-e4cc-4d25-9f42-2182928d43e4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.456120 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"178d7163-e4cc-4d25-9f42-2182928d43e4","Type":"ContainerDied","Data":"941de66ab894a71ec706670847807fc7ee1a23b1f03cc5a89f3b17afb5224c04"} Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.456591 4930 scope.go:117] "RemoveContainer" containerID="c6936094e864674206cc13c48094456a401ee0af5803bd17c934cdc8e261e538" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.456421 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.463267 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" event={"ID":"7548398c-45a9-4165-81fe-3f15f68dc161","Type":"ContainerDied","Data":"c4d8a90d361b52f4759606081f4ed85a048543f130355db7bbb67e82d25c4cf6"} Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.463375 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c4d8a90d361b52f4759606081f4ed85a048543f130355db7bbb67e82d25c4cf6" Mar 13 09:36:41 crc kubenswrapper[4930]: E0313 09:36:41.516785 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2deba84216225a676e1f124c0743a175b8a64979374dc1434ee67b9e91ef31ad is running failed: container process not found" containerID="2deba84216225a676e1f124c0743a175b8a64979374dc1434ee67b9e91ef31ad" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 09:36:41 crc kubenswrapper[4930]: E0313 09:36:41.518948 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2deba84216225a676e1f124c0743a175b8a64979374dc1434ee67b9e91ef31ad is running failed: container process not found" containerID="2deba84216225a676e1f124c0743a175b8a64979374dc1434ee67b9e91ef31ad" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 09:36:41 crc kubenswrapper[4930]: E0313 09:36:41.519183 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2deba84216225a676e1f124c0743a175b8a64979374dc1434ee67b9e91ef31ad is running failed: container process not found" containerID="2deba84216225a676e1f124c0743a175b8a64979374dc1434ee67b9e91ef31ad" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 09:36:41 crc kubenswrapper[4930]: E0313 09:36:41.519225 4930 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 2deba84216225a676e1f124c0743a175b8a64979374dc1434ee67b9e91ef31ad is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-62v7d" podUID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" containerName="registry-server" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.720698 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.748859 4930 scope.go:117] "RemoveContainer" containerID="b4d57afb3f9f4b4daa1177a0a7464e8a09b60d23e099df13c0deae7d6fbc4319" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.760500 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.782092 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.795841 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:36:41 crc kubenswrapper[4930]: E0313 09:36:41.796373 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7548398c-45a9-4165-81fe-3f15f68dc161" containerName="heat-cfnapi" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.796392 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="7548398c-45a9-4165-81fe-3f15f68dc161" containerName="heat-cfnapi" Mar 13 09:36:41 crc kubenswrapper[4930]: E0313 09:36:41.796409 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178d7163-e4cc-4d25-9f42-2182928d43e4" containerName="glance-log" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.796415 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="178d7163-e4cc-4d25-9f42-2182928d43e4" containerName="glance-log" Mar 13 09:36:41 crc kubenswrapper[4930]: E0313 09:36:41.796444 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="178d7163-e4cc-4d25-9f42-2182928d43e4" containerName="glance-httpd" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.796450 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="178d7163-e4cc-4d25-9f42-2182928d43e4" containerName="glance-httpd" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.796661 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="178d7163-e4cc-4d25-9f42-2182928d43e4" containerName="glance-httpd" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.796679 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="178d7163-e4cc-4d25-9f42-2182928d43e4" containerName="glance-log" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.796696 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="7548398c-45a9-4165-81fe-3f15f68dc161" containerName="heat-cfnapi" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.797961 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.805771 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.805958 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.854001 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.880727 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-config-data\") pod \"7548398c-45a9-4165-81fe-3f15f68dc161\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.881112 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-combined-ca-bundle\") pod \"7548398c-45a9-4165-81fe-3f15f68dc161\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.881164 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-config-data-custom\") pod \"7548398c-45a9-4165-81fe-3f15f68dc161\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.881288 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tq9tq\" (UniqueName: \"kubernetes.io/projected/7548398c-45a9-4165-81fe-3f15f68dc161-kube-api-access-tq9tq\") pod \"7548398c-45a9-4165-81fe-3f15f68dc161\" (UID: \"7548398c-45a9-4165-81fe-3f15f68dc161\") " Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.881688 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65aba82c-e055-4099-845f-72257d9295f0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.881727 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65aba82c-e055-4099-845f-72257d9295f0-logs\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.881777 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/65aba82c-e055-4099-845f-72257d9295f0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.881835 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65aba82c-e055-4099-845f-72257d9295f0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.881959 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.881984 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65aba82c-e055-4099-845f-72257d9295f0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.882013 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h8rx\" (UniqueName: \"kubernetes.io/projected/65aba82c-e055-4099-845f-72257d9295f0-kube-api-access-8h8rx\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.882054 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65aba82c-e055-4099-845f-72257d9295f0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.891166 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7548398c-45a9-4165-81fe-3f15f68dc161" (UID: "7548398c-45a9-4165-81fe-3f15f68dc161"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.891947 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7548398c-45a9-4165-81fe-3f15f68dc161-kube-api-access-tq9tq" (OuterVolumeSpecName: "kube-api-access-tq9tq") pod "7548398c-45a9-4165-81fe-3f15f68dc161" (UID: "7548398c-45a9-4165-81fe-3f15f68dc161"). InnerVolumeSpecName "kube-api-access-tq9tq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.939773 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7548398c-45a9-4165-81fe-3f15f68dc161" (UID: "7548398c-45a9-4165-81fe-3f15f68dc161"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.984067 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65aba82c-e055-4099-845f-72257d9295f0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.984219 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.984245 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65aba82c-e055-4099-845f-72257d9295f0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.984275 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h8rx\" (UniqueName: \"kubernetes.io/projected/65aba82c-e055-4099-845f-72257d9295f0-kube-api-access-8h8rx\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.984318 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65aba82c-e055-4099-845f-72257d9295f0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.984417 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65aba82c-e055-4099-845f-72257d9295f0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.984460 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65aba82c-e055-4099-845f-72257d9295f0-logs\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.984498 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/65aba82c-e055-4099-845f-72257d9295f0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.984564 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.984574 4930 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.984583 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tq9tq\" (UniqueName: \"kubernetes.io/projected/7548398c-45a9-4165-81fe-3f15f68dc161-kube-api-access-tq9tq\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.985003 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/65aba82c-e055-4099-845f-72257d9295f0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.996398 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.996719 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/a4f4242b5f3665838ee27f4f8b1e1823b22b221bc7d207651abbb8d37832373d/globalmount\"" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:41 crc kubenswrapper[4930]: I0313 09:36:41.998962 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65aba82c-e055-4099-845f-72257d9295f0-logs\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.001491 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/65aba82c-e055-4099-845f-72257d9295f0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.002353 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65aba82c-e055-4099-845f-72257d9295f0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.019635 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65aba82c-e055-4099-845f-72257d9295f0-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.020871 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/65aba82c-e055-4099-845f-72257d9295f0-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.027335 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h8rx\" (UniqueName: \"kubernetes.io/projected/65aba82c-e055-4099-845f-72257d9295f0-kube-api-access-8h8rx\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.046828 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="178d7163-e4cc-4d25-9f42-2182928d43e4" path="/var/lib/kubelet/pods/178d7163-e4cc-4d25-9f42-2182928d43e4/volumes" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.133605 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-config-data" (OuterVolumeSpecName: "config-data") pod "7548398c-45a9-4165-81fe-3f15f68dc161" (UID: "7548398c-45a9-4165-81fe-3f15f68dc161"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.189540 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7548398c-45a9-4165-81fe-3f15f68dc161-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.190568 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-893414a0-cbec-42bd-a89f-5c787585718f\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-893414a0-cbec-42bd-a89f-5c787585718f\") pod \"glance-default-internal-api-0\" (UID: \"65aba82c-e055-4099-845f-72257d9295f0\") " pod="openstack/glance-default-internal-api-0" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.202306 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.215949 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.248188 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.268145 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.298478 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-ovsdbserver-sb\") pod \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.298954 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nrpc\" (UniqueName: \"kubernetes.io/projected/d95152a3-5e16-464b-9785-88e629cc31f9-kube-api-access-8nrpc\") pod \"d95152a3-5e16-464b-9785-88e629cc31f9\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.299027 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x9h6b\" (UniqueName: \"kubernetes.io/projected/c765eaa6-eb16-4c09-b845-430cfd29e6d6-kube-api-access-x9h6b\") pod \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.299090 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-utilities\") pod \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\" (UID: \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.301806 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-utilities" (OuterVolumeSpecName: "utilities") pod "0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" (UID: "0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.306622 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"d95152a3-5e16-464b-9785-88e629cc31f9\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.316694 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d95152a3-5e16-464b-9785-88e629cc31f9-kube-api-access-8nrpc" (OuterVolumeSpecName: "kube-api-access-8nrpc") pod "d95152a3-5e16-464b-9785-88e629cc31f9" (UID: "d95152a3-5e16-464b-9785-88e629cc31f9"). InnerVolumeSpecName "kube-api-access-8nrpc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.340098 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-catalog-content\") pod \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\" (UID: \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.353762 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-dns-swift-storage-0\") pod \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.353875 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-public-tls-certs\") pod \"d95152a3-5e16-464b-9785-88e629cc31f9\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.353974 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-config\") pod \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.354040 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d95152a3-5e16-464b-9785-88e629cc31f9-logs\") pod \"d95152a3-5e16-464b-9785-88e629cc31f9\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.354969 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7sw8\" (UniqueName: \"kubernetes.io/projected/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-kube-api-access-j7sw8\") pod \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\" (UID: \"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.355041 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ln75n\" (UniqueName: \"kubernetes.io/projected/409ff6c8-9776-400c-9205-58658655a6a2-kube-api-access-ln75n\") pod \"409ff6c8-9776-400c-9205-58658655a6a2\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.355108 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-ovsdbserver-nb\") pod \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.355197 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-config-data\") pod \"409ff6c8-9776-400c-9205-58658655a6a2\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.355271 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-scripts\") pod \"d95152a3-5e16-464b-9785-88e629cc31f9\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.355303 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-config-data\") pod \"d95152a3-5e16-464b-9785-88e629cc31f9\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.355330 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-config-data-custom\") pod \"409ff6c8-9776-400c-9205-58658655a6a2\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.355365 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d95152a3-5e16-464b-9785-88e629cc31f9-httpd-run\") pod \"d95152a3-5e16-464b-9785-88e629cc31f9\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.355423 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-combined-ca-bundle\") pod \"409ff6c8-9776-400c-9205-58658655a6a2\" (UID: \"409ff6c8-9776-400c-9205-58658655a6a2\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.355537 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-dns-svc\") pod \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\" (UID: \"c765eaa6-eb16-4c09-b845-430cfd29e6d6\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.355566 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-combined-ca-bundle\") pod \"d95152a3-5e16-464b-9785-88e629cc31f9\" (UID: \"d95152a3-5e16-464b-9785-88e629cc31f9\") " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.356911 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nrpc\" (UniqueName: \"kubernetes.io/projected/d95152a3-5e16-464b-9785-88e629cc31f9-kube-api-access-8nrpc\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.356930 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.366298 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c765eaa6-eb16-4c09-b845-430cfd29e6d6-kube-api-access-x9h6b" (OuterVolumeSpecName: "kube-api-access-x9h6b") pod "c765eaa6-eb16-4c09-b845-430cfd29e6d6" (UID: "c765eaa6-eb16-4c09-b845-430cfd29e6d6"). InnerVolumeSpecName "kube-api-access-x9h6b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.367535 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d95152a3-5e16-464b-9785-88e629cc31f9-logs" (OuterVolumeSpecName: "logs") pod "d95152a3-5e16-464b-9785-88e629cc31f9" (UID: "d95152a3-5e16-464b-9785-88e629cc31f9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.372218 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d95152a3-5e16-464b-9785-88e629cc31f9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d95152a3-5e16-464b-9785-88e629cc31f9" (UID: "d95152a3-5e16-464b-9785-88e629cc31f9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.395918 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/409ff6c8-9776-400c-9205-58658655a6a2-kube-api-access-ln75n" (OuterVolumeSpecName: "kube-api-access-ln75n") pod "409ff6c8-9776-400c-9205-58658655a6a2" (UID: "409ff6c8-9776-400c-9205-58658655a6a2"). InnerVolumeSpecName "kube-api-access-ln75n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.398329 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "409ff6c8-9776-400c-9205-58658655a6a2" (UID: "409ff6c8-9776-400c-9205-58658655a6a2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.401289 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-scripts" (OuterVolumeSpecName: "scripts") pod "d95152a3-5e16-464b-9785-88e629cc31f9" (UID: "d95152a3-5e16-464b-9785-88e629cc31f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.419235 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.439489 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958" (OuterVolumeSpecName: "glance") pod "d95152a3-5e16-464b-9785-88e629cc31f9" (UID: "d95152a3-5e16-464b-9785-88e629cc31f9"). InnerVolumeSpecName "pvc-e6a7410f-e5ed-48ef-993c-0deb96840958". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.442312 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" (UID: "0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.450908 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-kube-api-access-j7sw8" (OuterVolumeSpecName: "kube-api-access-j7sw8") pod "0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" (UID: "0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929"). InnerVolumeSpecName "kube-api-access-j7sw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.463331 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x9h6b\" (UniqueName: \"kubernetes.io/projected/c765eaa6-eb16-4c09-b845-430cfd29e6d6-kube-api-access-x9h6b\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.463384 4930 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") on node \"crc\" " Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.463403 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.463417 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d95152a3-5e16-464b-9785-88e629cc31f9-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.463445 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7sw8\" (UniqueName: \"kubernetes.io/projected/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929-kube-api-access-j7sw8\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.463461 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ln75n\" (UniqueName: \"kubernetes.io/projected/409ff6c8-9776-400c-9205-58658655a6a2-kube-api-access-ln75n\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.463472 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.463485 4930 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.463495 4930 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d95152a3-5e16-464b-9785-88e629cc31f9-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.515043 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-558f5d58ff-jkrpd" event={"ID":"409ff6c8-9776-400c-9205-58658655a6a2","Type":"ContainerDied","Data":"00865374bebca57b67f00c191ddb576709c34011519e45d5222bf35ac1ef1c48"} Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.515102 4930 scope.go:117] "RemoveContainer" containerID="d797a82022c47a61191ce267b77101d6c6e4ce50ad8932bd81c8d2086452fbeb" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.515288 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-558f5d58ff-jkrpd" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.528118 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-62v7d" event={"ID":"0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929","Type":"ContainerDied","Data":"f6d2a71e9f06179a3a119543c8e02d642bd682240e2c1907d9a0b15f1d3b8545"} Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.528197 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-62v7d" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.537203 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"9b1d8908-3b46-4270-b04c-6ad122632782","Type":"ContainerStarted","Data":"21ff957f933b897024a03561439978776eb058157773c388ffd18deaa4792479"} Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.563169 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.563830 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-7wx75" event={"ID":"c765eaa6-eb16-4c09-b845-430cfd29e6d6","Type":"ContainerDied","Data":"b63c6918069f1a6e5d341ac72e2774ff2e089353f795668beb943fe8bb147485"} Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.572776 4930 scope.go:117] "RemoveContainer" containerID="2deba84216225a676e1f124c0743a175b8a64979374dc1434ee67b9e91ef31ad" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.581811 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-f4dfb5f5b-pw8zp" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.583672 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.585131 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"d95152a3-5e16-464b-9785-88e629cc31f9","Type":"ContainerDied","Data":"a54fd8e66dc7f4c2b59d1ec89408b71f3a5fdfaa5eac229d6321b224230d254e"} Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.608057 4930 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.608387 4930 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e6a7410f-e5ed-48ef-993c-0deb96840958" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958") on node "crc" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.617987 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.165709478 podStartE2EDuration="24.617962753s" podCreationTimestamp="2026-03-13 09:36:18 +0000 UTC" firstStartedPulling="2026-03-13 09:36:19.38860095 +0000 UTC m=+1420.138515627" lastFinishedPulling="2026-03-13 09:36:40.840854225 +0000 UTC m=+1441.590768902" observedRunningTime="2026-03-13 09:36:42.564918685 +0000 UTC m=+1443.314833362" watchObservedRunningTime="2026-03-13 09:36:42.617962753 +0000 UTC m=+1443.367877430" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.620226 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c765eaa6-eb16-4c09-b845-430cfd29e6d6" (UID: "c765eaa6-eb16-4c09-b845-430cfd29e6d6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.645768 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "409ff6c8-9776-400c-9205-58658655a6a2" (UID: "409ff6c8-9776-400c-9205-58658655a6a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.659967 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-config" (OuterVolumeSpecName: "config") pod "c765eaa6-eb16-4c09-b845-430cfd29e6d6" (UID: "c765eaa6-eb16-4c09-b845-430cfd29e6d6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.661818 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-62v7d"] Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.664229 4930 scope.go:117] "RemoveContainer" containerID="d32a14d102f8fdb0a134eac03d3661cbe076acce5ba1643171c5c8f9157b177d" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.669542 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.669577 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.669641 4930 reconciler_common.go:293] "Volume detached for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.669663 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.671185 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d95152a3-5e16-464b-9785-88e629cc31f9" (UID: "d95152a3-5e16-464b-9785-88e629cc31f9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.687217 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-62v7d"] Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.700538 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-f4dfb5f5b-pw8zp"] Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.700883 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-config-data" (OuterVolumeSpecName: "config-data") pod "d95152a3-5e16-464b-9785-88e629cc31f9" (UID: "d95152a3-5e16-464b-9785-88e629cc31f9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.703068 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c765eaa6-eb16-4c09-b845-430cfd29e6d6" (UID: "c765eaa6-eb16-4c09-b845-430cfd29e6d6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.716164 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c765eaa6-eb16-4c09-b845-430cfd29e6d6" (UID: "c765eaa6-eb16-4c09-b845-430cfd29e6d6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.716504 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d95152a3-5e16-464b-9785-88e629cc31f9" (UID: "d95152a3-5e16-464b-9785-88e629cc31f9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.722651 4930 scope.go:117] "RemoveContainer" containerID="6028c0d550eee00bf31309724244e206fe8871eee28e65a2e6ab2933370f8753" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.724656 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-f4dfb5f5b-pw8zp"] Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.735103 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c765eaa6-eb16-4c09-b845-430cfd29e6d6" (UID: "c765eaa6-eb16-4c09-b845-430cfd29e6d6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.737013 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-config-data" (OuterVolumeSpecName: "config-data") pod "409ff6c8-9776-400c-9205-58658655a6a2" (UID: "409ff6c8-9776-400c-9205-58658655a6a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.756295 4930 scope.go:117] "RemoveContainer" containerID="ff2905cd51a663f0a9d93f7c4daaf6804a2402ddf3b658878cb899db630b23cd" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.771484 4930 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.771539 4930 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.771552 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.771563 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/409ff6c8-9776-400c-9205-58658655a6a2-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.771575 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.771583 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d95152a3-5e16-464b-9785-88e629cc31f9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.771591 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c765eaa6-eb16-4c09-b845-430cfd29e6d6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.784711 4930 scope.go:117] "RemoveContainer" containerID="3472ada2eb325f5c27cfe7c71ef733160d23d45e2fdf21f36a61234fc7167d3d" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.811074 4930 scope.go:117] "RemoveContainer" containerID="29fb594a8475f74faf9b69578f86c18f499f27eb837df8c5196f92b52408d168" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.858388 4930 scope.go:117] "RemoveContainer" containerID="995636287c1ceecdc41a41988d0a3367c860c034253f4dffe261994bdfc30a8f" Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.874279 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-558f5d58ff-jkrpd"] Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.885012 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-558f5d58ff-jkrpd"] Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.905062 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-77757997f5-wlmc4"] Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.919419 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7wx75"] Mar 13 09:36:42 crc kubenswrapper[4930]: I0313 09:36:42.927820 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-7wx75"] Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.164485 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.223640 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.242698 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:36:43 crc kubenswrapper[4930]: E0313 09:36:43.243260 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c765eaa6-eb16-4c09-b845-430cfd29e6d6" containerName="init" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243273 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c765eaa6-eb16-4c09-b845-430cfd29e6d6" containerName="init" Mar 13 09:36:43 crc kubenswrapper[4930]: E0313 09:36:43.243290 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95152a3-5e16-464b-9785-88e629cc31f9" containerName="glance-httpd" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243297 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95152a3-5e16-464b-9785-88e629cc31f9" containerName="glance-httpd" Mar 13 09:36:43 crc kubenswrapper[4930]: E0313 09:36:43.243309 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" containerName="registry-server" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243316 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" containerName="registry-server" Mar 13 09:36:43 crc kubenswrapper[4930]: E0313 09:36:43.243326 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" containerName="extract-utilities" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243334 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" containerName="extract-utilities" Mar 13 09:36:43 crc kubenswrapper[4930]: E0313 09:36:43.243375 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="409ff6c8-9776-400c-9205-58658655a6a2" containerName="heat-api" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243382 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="409ff6c8-9776-400c-9205-58658655a6a2" containerName="heat-api" Mar 13 09:36:43 crc kubenswrapper[4930]: E0313 09:36:43.243393 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" containerName="extract-content" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243398 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" containerName="extract-content" Mar 13 09:36:43 crc kubenswrapper[4930]: E0313 09:36:43.243412 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c765eaa6-eb16-4c09-b845-430cfd29e6d6" containerName="dnsmasq-dns" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243418 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c765eaa6-eb16-4c09-b845-430cfd29e6d6" containerName="dnsmasq-dns" Mar 13 09:36:43 crc kubenswrapper[4930]: E0313 09:36:43.243444 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d95152a3-5e16-464b-9785-88e629cc31f9" containerName="glance-log" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243452 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d95152a3-5e16-464b-9785-88e629cc31f9" containerName="glance-log" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243663 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95152a3-5e16-464b-9785-88e629cc31f9" containerName="glance-log" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243679 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="409ff6c8-9776-400c-9205-58658655a6a2" containerName="heat-api" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243689 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" containerName="registry-server" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243706 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d95152a3-5e16-464b-9785-88e629cc31f9" containerName="glance-httpd" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.243715 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c765eaa6-eb16-4c09-b845-430cfd29e6d6" containerName="dnsmasq-dns" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.245238 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.247602 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.247750 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.253403 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.285845 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0055ea30-43e8-4914-a0b3-19b611e2f82f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.285908 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.285983 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0055ea30-43e8-4914-a0b3-19b611e2f82f-scripts\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.286067 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0055ea30-43e8-4914-a0b3-19b611e2f82f-logs\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.286178 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0055ea30-43e8-4914-a0b3-19b611e2f82f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.286253 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0055ea30-43e8-4914-a0b3-19b611e2f82f-config-data\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.286287 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbvw5\" (UniqueName: \"kubernetes.io/projected/0055ea30-43e8-4914-a0b3-19b611e2f82f-kube-api-access-bbvw5\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.286450 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0055ea30-43e8-4914-a0b3-19b611e2f82f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.390348 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0055ea30-43e8-4914-a0b3-19b611e2f82f-logs\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.390605 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0055ea30-43e8-4914-a0b3-19b611e2f82f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.391216 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0055ea30-43e8-4914-a0b3-19b611e2f82f-logs\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.391269 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/0055ea30-43e8-4914-a0b3-19b611e2f82f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.391524 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0055ea30-43e8-4914-a0b3-19b611e2f82f-config-data\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.391589 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbvw5\" (UniqueName: \"kubernetes.io/projected/0055ea30-43e8-4914-a0b3-19b611e2f82f-kube-api-access-bbvw5\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.392005 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0055ea30-43e8-4914-a0b3-19b611e2f82f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.392683 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0055ea30-43e8-4914-a0b3-19b611e2f82f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.392757 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.392887 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0055ea30-43e8-4914-a0b3-19b611e2f82f-scripts\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.397154 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0055ea30-43e8-4914-a0b3-19b611e2f82f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.397893 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.397927 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/99223c607948bed00b9fe5109399df143422767dfa8a32b71d4b371c3748c0ce/globalmount\"" pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.399039 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0055ea30-43e8-4914-a0b3-19b611e2f82f-config-data\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.400980 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0055ea30-43e8-4914-a0b3-19b611e2f82f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.409722 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0055ea30-43e8-4914-a0b3-19b611e2f82f-scripts\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.414161 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbvw5\" (UniqueName: \"kubernetes.io/projected/0055ea30-43e8-4914-a0b3-19b611e2f82f-kube-api-access-bbvw5\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.522344 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e6a7410f-e5ed-48ef-993c-0deb96840958\") pod \"glance-default-external-api-0\" (UID: \"0055ea30-43e8-4914-a0b3-19b611e2f82f\") " pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.589277 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.599377 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 13 09:36:43 crc kubenswrapper[4930]: W0313 09:36:43.605816 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec86dec2_eb5a_497e_bba2_d1b96d80c65c.slice/crio-c10436c49682d1aeaf2220e08e3f1c75d1e793acb430bb10d97f096c531adb4e WatchSource:0}: Error finding container c10436c49682d1aeaf2220e08e3f1c75d1e793acb430bb10d97f096c531adb4e: Status 404 returned error can't find the container with id c10436c49682d1aeaf2220e08e3f1c75d1e793acb430bb10d97f096c531adb4e Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.641714 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e5091765-b1c8-43be-9579-d503faf84b6e","Type":"ContainerStarted","Data":"c3b574ad5214038a41c7a4cbd2590943f742ff6b249f2702947402114174d225"} Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.646107 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7889c489fb-hj2f9"] Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.647044 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7889c489fb-hj2f9" event={"ID":"304ba316-73ff-420c-af72-945f29aa5e2e","Type":"ContainerStarted","Data":"fc6e2d22a8fceecf478d19c8a447a42ce80ddf3c1589ffb1bf9099ae6abbe3bd"} Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.663173 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77757997f5-wlmc4" event={"ID":"0a8d2ceb-f27f-4e71-9c58-64cb813bd385","Type":"ContainerStarted","Data":"1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59"} Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.663218 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77757997f5-wlmc4" event={"ID":"0a8d2ceb-f27f-4e71-9c58-64cb813bd385","Type":"ContainerStarted","Data":"a20c6ae78f99aae409e30b2ea2567054f5a9075cf0a6f7da5f6ac13dcc528aac"} Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.664273 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.694482 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-qbssl"] Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.769815 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-b5f3-account-create-update-9zsnn"] Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.811631 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-122f-account-create-update-9988g"] Mar 13 09:36:43 crc kubenswrapper[4930]: W0313 09:36:43.826659 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a92113b_21a5_4db3_9708_589fc0aa57e8.slice/crio-7fb40d52130773d34237fbfe656f5ada4da1abeabb2c2598579ce8bcf4e5c164 WatchSource:0}: Error finding container 7fb40d52130773d34237fbfe656f5ada4da1abeabb2c2598579ce8bcf4e5c164: Status 404 returned error can't find the container with id 7fb40d52130773d34237fbfe656f5ada4da1abeabb2c2598579ce8bcf4e5c164 Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.829668 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-xxmsm"] Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.872765 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.879420 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-77757997f5-wlmc4" podStartSLOduration=9.879395813 podStartE2EDuration="9.879395813s" podCreationTimestamp="2026-03-13 09:36:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:43.728076482 +0000 UTC m=+1444.477991169" watchObservedRunningTime="2026-03-13 09:36:43.879395813 +0000 UTC m=+1444.629310490" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.880104 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-lzrnq"] Mar 13 09:36:43 crc kubenswrapper[4930]: W0313 09:36:43.895880 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1a0aa4e_26e8_4ff5_9850_df0535655e9e.slice/crio-01a84cac61dddc07677cb61c72f2195a66659952ef9c47f94e38afb874c51f8e WatchSource:0}: Error finding container 01a84cac61dddc07677cb61c72f2195a66659952ef9c47f94e38afb874c51f8e: Status 404 returned error can't find the container with id 01a84cac61dddc07677cb61c72f2195a66659952ef9c47f94e38afb874c51f8e Mar 13 09:36:43 crc kubenswrapper[4930]: W0313 09:36:43.898182 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod17c2545f_f1f1_4c0a_a689_73202f48ddf6.slice/crio-5086aacf2d28d8b560aa097b6a53ab3b7d62cb536a3970c14d6255d4055fc66d WatchSource:0}: Error finding container 5086aacf2d28d8b560aa097b6a53ab3b7d62cb536a3970c14d6255d4055fc66d: Status 404 returned error can't find the container with id 5086aacf2d28d8b560aa097b6a53ab3b7d62cb536a3970c14d6255d4055fc66d Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.924166 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.936592 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 13 09:36:43 crc kubenswrapper[4930]: I0313 09:36:43.965287 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-c6b696d47-c76xv"] Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.051574 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929" path="/var/lib/kubelet/pods/0f9507e9-9cda-4c6c-bb9f-c4b7ebf92929/volumes" Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.077539 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="409ff6c8-9776-400c-9205-58658655a6a2" path="/var/lib/kubelet/pods/409ff6c8-9776-400c-9205-58658655a6a2/volumes" Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.079517 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7548398c-45a9-4165-81fe-3f15f68dc161" path="/var/lib/kubelet/pods/7548398c-45a9-4165-81fe-3f15f68dc161/volumes" Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.081913 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c765eaa6-eb16-4c09-b845-430cfd29e6d6" path="/var/lib/kubelet/pods/c765eaa6-eb16-4c09-b845-430cfd29e6d6/volumes" Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.084211 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d95152a3-5e16-464b-9785-88e629cc31f9" path="/var/lib/kubelet/pods/d95152a3-5e16-464b-9785-88e629cc31f9/volumes" Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.090943 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-7e6b-account-create-update-tgmkk"] Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.091332 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.091579 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-86b9fc57f7-vjhx9"] Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.091606 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-66d7f6c554-7d4pc"] Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.205292 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 13 09:36:44 crc kubenswrapper[4930]: W0313 09:36:44.241350 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65aba82c_e055_4099_845f_72257d9295f0.slice/crio-6ad0a4c125dedd2e47b350ee7c87bdcb061387a1bf809564886f995be4b55bfe WatchSource:0}: Error finding container 6ad0a4c125dedd2e47b350ee7c87bdcb061387a1bf809564886f995be4b55bfe: Status 404 returned error can't find the container with id 6ad0a4c125dedd2e47b350ee7c87bdcb061387a1bf809564886f995be4b55bfe Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.287212 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.572906 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.739293 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xxmsm" event={"ID":"b3c2e237-7f4c-48a2-8a8d-f5838192decc","Type":"ContainerStarted","Data":"3dbe2bd59a2ee942e1a638ba812901c78eaa63afb2fd8213944b97ef7b145488"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.740514 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"65aba82c-e055-4099-845f-72257d9295f0","Type":"ContainerStarted","Data":"6ad0a4c125dedd2e47b350ee7c87bdcb061387a1bf809564886f995be4b55bfe"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.742144 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d82ac246-17a4-4f81-acce-70aaa7d8f350","Type":"ContainerStarted","Data":"82212ef0f69d40965d51b48fbde952d4247c9da3152964fe258e7c07693b076c"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.743931 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lzrnq" event={"ID":"9e12cb85-3947-4c60-96a6-8ac07f5e63e8","Type":"ContainerStarted","Data":"3b38c16cad245857ce3b91e4f3cb3027749ec8a011d0d5ea30a2e543f318b8db"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.747336 4930 generic.go:334] "Generic (PLEG): container finished" podID="304ba316-73ff-420c-af72-945f29aa5e2e" containerID="99361870f3e59d579a74a27abb46429b1fb6039c1077f8a74a1db641cc94caf9" exitCode=1 Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.747381 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7889c489fb-hj2f9" event={"ID":"304ba316-73ff-420c-af72-945f29aa5e2e","Type":"ContainerDied","Data":"99361870f3e59d579a74a27abb46429b1fb6039c1077f8a74a1db641cc94caf9"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.748380 4930 scope.go:117] "RemoveContainer" containerID="99361870f3e59d579a74a27abb46429b1fb6039c1077f8a74a1db641cc94caf9" Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.749382 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0055ea30-43e8-4914-a0b3-19b611e2f82f","Type":"ContainerStarted","Data":"1fbda3f3885c41231b0ab48f92786db82a3484d0d5d5e6400a564ae77fd0e5a2"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.754514 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" event={"ID":"17c2545f-f1f1-4c0a-a689-73202f48ddf6","Type":"ContainerStarted","Data":"5086aacf2d28d8b560aa097b6a53ab3b7d62cb536a3970c14d6255d4055fc66d"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.758650 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-122f-account-create-update-9988g" event={"ID":"f5942b4a-ef6d-40ca-b67f-e7965513004d","Type":"ContainerStarted","Data":"fc112d9419344d327e1c8c8a6349dd19da83a7b1e827b7fef330b29325fea0da"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.758697 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-122f-account-create-update-9988g" event={"ID":"f5942b4a-ef6d-40ca-b67f-e7965513004d","Type":"ContainerStarted","Data":"7393deb2540084be598480143ce606079bfc371b73bf5d92f7688fe4b8364691"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.765931 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" event={"ID":"4a15fab1-c914-4555-813a-ab9552b29a22","Type":"ContainerStarted","Data":"98653f43710ccbbef2f16818243db0fd693465f968e3d9bd3026ef6ae078899d"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.783950 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" event={"ID":"e1a0aa4e-26e8-4ff5-9850-df0535655e9e","Type":"ContainerStarted","Data":"01a84cac61dddc07677cb61c72f2195a66659952ef9c47f94e38afb874c51f8e"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.796033 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-c6b696d47-c76xv" event={"ID":"f0868712-845b-4803-b7a7-44aaa23ad284","Type":"ContainerStarted","Data":"d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.796074 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-c6b696d47-c76xv" event={"ID":"f0868712-845b-4803-b7a7-44aaa23ad284","Type":"ContainerStarted","Data":"47b3c8d3318c0e56add86aca51d17f8bc3f46e4536e054a342375d27f53241f4"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.797413 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.800913 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-122f-account-create-update-9988g" podStartSLOduration=5.800891765 podStartE2EDuration="5.800891765s" podCreationTimestamp="2026-03-13 09:36:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:44.787995608 +0000 UTC m=+1445.537910285" watchObservedRunningTime="2026-03-13 09:36:44.800891765 +0000 UTC m=+1445.550806442" Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.813120 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qbssl" event={"ID":"ec86dec2-eb5a-497e-bba2-d1b96d80c65c","Type":"ContainerStarted","Data":"c10436c49682d1aeaf2220e08e3f1c75d1e793acb430bb10d97f096c531adb4e"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.820804 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b5f3-account-create-update-9zsnn" event={"ID":"4a92113b-21a5-4db3-9708-589fc0aa57e8","Type":"ContainerStarted","Data":"7fb40d52130773d34237fbfe656f5ada4da1abeabb2c2598579ce8bcf4e5c164"} Mar 13 09:36:44 crc kubenswrapper[4930]: I0313 09:36:44.837377 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-c6b696d47-c76xv" podStartSLOduration=11.83735437 podStartE2EDuration="11.83735437s" podCreationTimestamp="2026-03-13 09:36:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:44.812756356 +0000 UTC m=+1445.562671033" watchObservedRunningTime="2026-03-13 09:36:44.83735437 +0000 UTC m=+1445.587269047" Mar 13 09:36:45 crc kubenswrapper[4930]: I0313 09:36:45.831507 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b5f3-account-create-update-9zsnn" event={"ID":"4a92113b-21a5-4db3-9708-589fc0aa57e8","Type":"ContainerStarted","Data":"2fd9d220e0ef2e9c5e49e551039bcf3e92dafde12abc470bdf438161092241d1"} Mar 13 09:36:45 crc kubenswrapper[4930]: I0313 09:36:45.834938 4930 generic.go:334] "Generic (PLEG): container finished" podID="f5942b4a-ef6d-40ca-b67f-e7965513004d" containerID="fc112d9419344d327e1c8c8a6349dd19da83a7b1e827b7fef330b29325fea0da" exitCode=0 Mar 13 09:36:45 crc kubenswrapper[4930]: I0313 09:36:45.834984 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-122f-account-create-update-9988g" event={"ID":"f5942b4a-ef6d-40ca-b67f-e7965513004d","Type":"ContainerDied","Data":"fc112d9419344d327e1c8c8a6349dd19da83a7b1e827b7fef330b29325fea0da"} Mar 13 09:36:45 crc kubenswrapper[4930]: I0313 09:36:45.840199 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" event={"ID":"4a15fab1-c914-4555-813a-ab9552b29a22","Type":"ContainerStarted","Data":"4b7b5aa15cd29fc57e05fac737b0fb42c5eb2c5caaab5c9fa05a061b5338ee86"} Mar 13 09:36:45 crc kubenswrapper[4930]: I0313 09:36:45.844311 4930 generic.go:334] "Generic (PLEG): container finished" podID="b3c2e237-7f4c-48a2-8a8d-f5838192decc" containerID="8eb6dd8aa8cf56e87cf75580b57bbd99dd58ff1bac13350056a5a725df6555b4" exitCode=0 Mar 13 09:36:45 crc kubenswrapper[4930]: I0313 09:36:45.844512 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xxmsm" event={"ID":"b3c2e237-7f4c-48a2-8a8d-f5838192decc","Type":"ContainerDied","Data":"8eb6dd8aa8cf56e87cf75580b57bbd99dd58ff1bac13350056a5a725df6555b4"} Mar 13 09:36:45 crc kubenswrapper[4930]: I0313 09:36:45.846786 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lzrnq" event={"ID":"9e12cb85-3947-4c60-96a6-8ac07f5e63e8","Type":"ContainerStarted","Data":"ddbb33d16390e104959af69b3bda2520dd6ea5437de7ac7f96f224286260373f"} Mar 13 09:36:45 crc kubenswrapper[4930]: I0313 09:36:45.850118 4930 generic.go:334] "Generic (PLEG): container finished" podID="ec86dec2-eb5a-497e-bba2-d1b96d80c65c" containerID="eb1cc95c17cab0807e9c358ba5604ad7f4577cbc6877f1993d79c9e79ea03e29" exitCode=0 Mar 13 09:36:45 crc kubenswrapper[4930]: I0313 09:36:45.850792 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qbssl" event={"ID":"ec86dec2-eb5a-497e-bba2-d1b96d80c65c","Type":"ContainerDied","Data":"eb1cc95c17cab0807e9c358ba5604ad7f4577cbc6877f1993d79c9e79ea03e29"} Mar 13 09:36:45 crc kubenswrapper[4930]: I0313 09:36:45.858305 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-b5f3-account-create-update-9zsnn" podStartSLOduration=7.858286824 podStartE2EDuration="7.858286824s" podCreationTimestamp="2026-03-13 09:36:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:45.846603188 +0000 UTC m=+1446.596517875" watchObservedRunningTime="2026-03-13 09:36:45.858286824 +0000 UTC m=+1446.608201501" Mar 13 09:36:45 crc kubenswrapper[4930]: I0313 09:36:45.905479 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" podStartSLOduration=7.905455048 podStartE2EDuration="7.905455048s" podCreationTimestamp="2026-03-13 09:36:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:45.896498454 +0000 UTC m=+1446.646413131" watchObservedRunningTime="2026-03-13 09:36:45.905455048 +0000 UTC m=+1446.655369725" Mar 13 09:36:45 crc kubenswrapper[4930]: I0313 09:36:45.958344 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-lzrnq" podStartSLOduration=7.958321512 podStartE2EDuration="7.958321512s" podCreationTimestamp="2026-03-13 09:36:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:45.946959635 +0000 UTC m=+1446.696874312" watchObservedRunningTime="2026-03-13 09:36:45.958321512 +0000 UTC m=+1446.708236189" Mar 13 09:36:47 crc kubenswrapper[4930]: I0313 09:36:47.330052 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bsl74" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerName="registry-server" probeResult="failure" output=< Mar 13 09:36:47 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:36:47 crc kubenswrapper[4930]: > Mar 13 09:36:47 crc kubenswrapper[4930]: I0313 09:36:47.899699 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-xxmsm" event={"ID":"b3c2e237-7f4c-48a2-8a8d-f5838192decc","Type":"ContainerDied","Data":"3dbe2bd59a2ee942e1a638ba812901c78eaa63afb2fd8213944b97ef7b145488"} Mar 13 09:36:47 crc kubenswrapper[4930]: I0313 09:36:47.900191 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dbe2bd59a2ee942e1a638ba812901c78eaa63afb2fd8213944b97ef7b145488" Mar 13 09:36:47 crc kubenswrapper[4930]: I0313 09:36:47.910351 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-122f-account-create-update-9988g" Mar 13 09:36:47 crc kubenswrapper[4930]: I0313 09:36:47.910896 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qbssl" Mar 13 09:36:47 crc kubenswrapper[4930]: I0313 09:36:47.910990 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" event={"ID":"e1a0aa4e-26e8-4ff5-9850-df0535655e9e","Type":"ContainerStarted","Data":"3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce"} Mar 13 09:36:47 crc kubenswrapper[4930]: I0313 09:36:47.911908 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:47 crc kubenswrapper[4930]: I0313 09:36:47.978141 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xxmsm" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.034801 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" podStartSLOduration=14.034776107 podStartE2EDuration="14.034776107s" podCreationTimestamp="2026-03-13 09:36:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:48.023771459 +0000 UTC m=+1448.773686136" watchObservedRunningTime="2026-03-13 09:36:48.034776107 +0000 UTC m=+1448.784690784" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.039669 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e5091765-b1c8-43be-9579-d503faf84b6e","Type":"ContainerStarted","Data":"0c5ac89688d9ba42713fc6bb044f42a88503026717ee271ac665ad3740a28374"} Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.039704 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d82ac246-17a4-4f81-acce-70aaa7d8f350","Type":"ContainerStarted","Data":"eec06ec7c79f1ef02d6b984c95a57a07ab5e252fc2714c479396bd92af8761ef"} Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.039717 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"65aba82c-e055-4099-845f-72257d9295f0","Type":"ContainerStarted","Data":"f8891c055ffd1522fe8162efaa3af08f3742fa6d6511ab01d5d43db8b53d0bf6"} Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.040369 4930 generic.go:334] "Generic (PLEG): container finished" podID="9e12cb85-3947-4c60-96a6-8ac07f5e63e8" containerID="ddbb33d16390e104959af69b3bda2520dd6ea5437de7ac7f96f224286260373f" exitCode=0 Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.040961 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lzrnq" event={"ID":"9e12cb85-3947-4c60-96a6-8ac07f5e63e8","Type":"ContainerDied","Data":"ddbb33d16390e104959af69b3bda2520dd6ea5437de7ac7f96f224286260373f"} Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.059246 4930 generic.go:334] "Generic (PLEG): container finished" podID="4a92113b-21a5-4db3-9708-589fc0aa57e8" containerID="2fd9d220e0ef2e9c5e49e551039bcf3e92dafde12abc470bdf438161092241d1" exitCode=0 Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.059331 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b5f3-account-create-update-9zsnn" event={"ID":"4a92113b-21a5-4db3-9708-589fc0aa57e8","Type":"ContainerDied","Data":"2fd9d220e0ef2e9c5e49e551039bcf3e92dafde12abc470bdf438161092241d1"} Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.077634 4930 generic.go:334] "Generic (PLEG): container finished" podID="4a15fab1-c914-4555-813a-ab9552b29a22" containerID="4b7b5aa15cd29fc57e05fac737b0fb42c5eb2c5caaab5c9fa05a061b5338ee86" exitCode=0 Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.077695 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" event={"ID":"4a15fab1-c914-4555-813a-ab9552b29a22","Type":"ContainerDied","Data":"4b7b5aa15cd29fc57e05fac737b0fb42c5eb2c5caaab5c9fa05a061b5338ee86"} Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.096613 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec86dec2-eb5a-497e-bba2-d1b96d80c65c-operator-scripts\") pod \"ec86dec2-eb5a-497e-bba2-d1b96d80c65c\" (UID: \"ec86dec2-eb5a-497e-bba2-d1b96d80c65c\") " Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.096863 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5942b4a-ef6d-40ca-b67f-e7965513004d-operator-scripts\") pod \"f5942b4a-ef6d-40ca-b67f-e7965513004d\" (UID: \"f5942b4a-ef6d-40ca-b67f-e7965513004d\") " Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.097176 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rsdt8\" (UniqueName: \"kubernetes.io/projected/f5942b4a-ef6d-40ca-b67f-e7965513004d-kube-api-access-rsdt8\") pod \"f5942b4a-ef6d-40ca-b67f-e7965513004d\" (UID: \"f5942b4a-ef6d-40ca-b67f-e7965513004d\") " Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.097273 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3c2e237-7f4c-48a2-8a8d-f5838192decc-operator-scripts\") pod \"b3c2e237-7f4c-48a2-8a8d-f5838192decc\" (UID: \"b3c2e237-7f4c-48a2-8a8d-f5838192decc\") " Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.097506 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f4ml6\" (UniqueName: \"kubernetes.io/projected/b3c2e237-7f4c-48a2-8a8d-f5838192decc-kube-api-access-f4ml6\") pod \"b3c2e237-7f4c-48a2-8a8d-f5838192decc\" (UID: \"b3c2e237-7f4c-48a2-8a8d-f5838192decc\") " Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.097625 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grt5l\" (UniqueName: \"kubernetes.io/projected/ec86dec2-eb5a-497e-bba2-d1b96d80c65c-kube-api-access-grt5l\") pod \"ec86dec2-eb5a-497e-bba2-d1b96d80c65c\" (UID: \"ec86dec2-eb5a-497e-bba2-d1b96d80c65c\") " Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.100072 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ec86dec2-eb5a-497e-bba2-d1b96d80c65c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ec86dec2-eb5a-497e-bba2-d1b96d80c65c" (UID: "ec86dec2-eb5a-497e-bba2-d1b96d80c65c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.101986 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3c2e237-7f4c-48a2-8a8d-f5838192decc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b3c2e237-7f4c-48a2-8a8d-f5838192decc" (UID: "b3c2e237-7f4c-48a2-8a8d-f5838192decc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.104487 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3c2e237-7f4c-48a2-8a8d-f5838192decc-kube-api-access-f4ml6" (OuterVolumeSpecName: "kube-api-access-f4ml6") pod "b3c2e237-7f4c-48a2-8a8d-f5838192decc" (UID: "b3c2e237-7f4c-48a2-8a8d-f5838192decc"). InnerVolumeSpecName "kube-api-access-f4ml6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.104910 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f5942b4a-ef6d-40ca-b67f-e7965513004d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f5942b4a-ef6d-40ca-b67f-e7965513004d" (UID: "f5942b4a-ef6d-40ca-b67f-e7965513004d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.106698 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ec86dec2-eb5a-497e-bba2-d1b96d80c65c-kube-api-access-grt5l" (OuterVolumeSpecName: "kube-api-access-grt5l") pod "ec86dec2-eb5a-497e-bba2-d1b96d80c65c" (UID: "ec86dec2-eb5a-497e-bba2-d1b96d80c65c"). InnerVolumeSpecName "kube-api-access-grt5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.107846 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5942b4a-ef6d-40ca-b67f-e7965513004d-kube-api-access-rsdt8" (OuterVolumeSpecName: "kube-api-access-rsdt8") pod "f5942b4a-ef6d-40ca-b67f-e7965513004d" (UID: "f5942b4a-ef6d-40ca-b67f-e7965513004d"). InnerVolumeSpecName "kube-api-access-rsdt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.108186 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-122f-account-create-update-9988g" event={"ID":"f5942b4a-ef6d-40ca-b67f-e7965513004d","Type":"ContainerDied","Data":"7393deb2540084be598480143ce606079bfc371b73bf5d92f7688fe4b8364691"} Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.108220 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7393deb2540084be598480143ce606079bfc371b73bf5d92f7688fe4b8364691" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.108323 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-122f-account-create-update-9988g" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.134877 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-qbssl" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.135009 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-qbssl" event={"ID":"ec86dec2-eb5a-497e-bba2-d1b96d80c65c","Type":"ContainerDied","Data":"c10436c49682d1aeaf2220e08e3f1c75d1e793acb430bb10d97f096c531adb4e"} Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.135037 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c10436c49682d1aeaf2220e08e3f1c75d1e793acb430bb10d97f096c531adb4e" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.146640 4930 generic.go:334] "Generic (PLEG): container finished" podID="304ba316-73ff-420c-af72-945f29aa5e2e" containerID="cdb8e49793e92f7f6013638a2391f36b891181793818f70e69c8578f03c9901c" exitCode=1 Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.146788 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7889c489fb-hj2f9" event={"ID":"304ba316-73ff-420c-af72-945f29aa5e2e","Type":"ContainerDied","Data":"cdb8e49793e92f7f6013638a2391f36b891181793818f70e69c8578f03c9901c"} Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.147392 4930 scope.go:117] "RemoveContainer" containerID="99361870f3e59d579a74a27abb46429b1fb6039c1077f8a74a1db641cc94caf9" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.151517 4930 scope.go:117] "RemoveContainer" containerID="cdb8e49793e92f7f6013638a2391f36b891181793818f70e69c8578f03c9901c" Mar 13 09:36:48 crc kubenswrapper[4930]: E0313 09:36:48.155456 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7889c489fb-hj2f9_openstack(304ba316-73ff-420c-af72-945f29aa5e2e)\"" pod="openstack/heat-api-7889c489fb-hj2f9" podUID="304ba316-73ff-420c-af72-945f29aa5e2e" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.161422 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0055ea30-43e8-4914-a0b3-19b611e2f82f","Type":"ContainerStarted","Data":"d82afc29e5f31b34f7ac8867d7c6b5a0f060833709f156a3303e86dce6c55fc6"} Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.209634 4930 generic.go:334] "Generic (PLEG): container finished" podID="17c2545f-f1f1-4c0a-a689-73202f48ddf6" containerID="141fe9f75845109674e45ebefab58e9b1f73e50ddbd88a7b327c71463595ae85" exitCode=1 Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.209693 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" event={"ID":"17c2545f-f1f1-4c0a-a689-73202f48ddf6","Type":"ContainerDied","Data":"141fe9f75845109674e45ebefab58e9b1f73e50ddbd88a7b327c71463595ae85"} Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.211223 4930 scope.go:117] "RemoveContainer" containerID="141fe9f75845109674e45ebefab58e9b1f73e50ddbd88a7b327c71463595ae85" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.212518 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rsdt8\" (UniqueName: \"kubernetes.io/projected/f5942b4a-ef6d-40ca-b67f-e7965513004d-kube-api-access-rsdt8\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.212549 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b3c2e237-7f4c-48a2-8a8d-f5838192decc-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.212559 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f4ml6\" (UniqueName: \"kubernetes.io/projected/b3c2e237-7f4c-48a2-8a8d-f5838192decc-kube-api-access-f4ml6\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.212569 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grt5l\" (UniqueName: \"kubernetes.io/projected/ec86dec2-eb5a-497e-bba2-d1b96d80c65c-kube-api-access-grt5l\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.212577 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ec86dec2-eb5a-497e-bba2-d1b96d80c65c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.212585 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f5942b4a-ef6d-40ca-b67f-e7965513004d-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:48 crc kubenswrapper[4930]: E0313 09:36:48.467813 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec86dec2_eb5a_497e_bba2_d1b96d80c65c.slice\": RecentStats: unable to find data in memory cache]" Mar 13 09:36:48 crc kubenswrapper[4930]: E0313 09:36:48.468040 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podec86dec2_eb5a_497e_bba2_d1b96d80c65c.slice\": RecentStats: unable to find data in memory cache]" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.508496 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.508555 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.519347 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:48 crc kubenswrapper[4930]: I0313 09:36:48.520245 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.227268 4930 generic.go:334] "Generic (PLEG): container finished" podID="17c2545f-f1f1-4c0a-a689-73202f48ddf6" containerID="ef9d3b908088bcbe51f586c23e75287474186c5ddf9f52b7ab287e61a55a9477" exitCode=1 Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.227601 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" event={"ID":"17c2545f-f1f1-4c0a-a689-73202f48ddf6","Type":"ContainerDied","Data":"ef9d3b908088bcbe51f586c23e75287474186c5ddf9f52b7ab287e61a55a9477"} Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.227637 4930 scope.go:117] "RemoveContainer" containerID="141fe9f75845109674e45ebefab58e9b1f73e50ddbd88a7b327c71463595ae85" Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.228232 4930 scope.go:117] "RemoveContainer" containerID="ef9d3b908088bcbe51f586c23e75287474186c5ddf9f52b7ab287e61a55a9477" Mar 13 09:36:49 crc kubenswrapper[4930]: E0313 09:36:49.228562 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-66d7f6c554-7d4pc_openstack(17c2545f-f1f1-4c0a-a689-73202f48ddf6)\"" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" podUID="17c2545f-f1f1-4c0a-a689-73202f48ddf6" Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.258464 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e5091765-b1c8-43be-9579-d503faf84b6e","Type":"ContainerStarted","Data":"8f167a7939d155e62597044e25b13583804dfa7aa665d5097cfc461951893973"} Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.262375 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"65aba82c-e055-4099-845f-72257d9295f0","Type":"ContainerStarted","Data":"26133727828b345676a04cac7d1a4c767678bed3e5fd98915aaf0ccc552e4595"} Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.267703 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d82ac246-17a4-4f81-acce-70aaa7d8f350","Type":"ContainerStarted","Data":"3ca96f8ad29d015617d7b897dce3c23513100a62779c5e71e4443f6ca246230b"} Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.268697 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.274049 4930 scope.go:117] "RemoveContainer" containerID="cdb8e49793e92f7f6013638a2391f36b891181793818f70e69c8578f03c9901c" Mar 13 09:36:49 crc kubenswrapper[4930]: E0313 09:36:49.274316 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7889c489fb-hj2f9_openstack(304ba316-73ff-420c-af72-945f29aa5e2e)\"" pod="openstack/heat-api-7889c489fb-hj2f9" podUID="304ba316-73ff-420c-af72-945f29aa5e2e" Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.288129 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"0055ea30-43e8-4914-a0b3-19b611e2f82f","Type":"ContainerStarted","Data":"0d904ceedd09dfb256e24c99df3d63539cfe2115ea4894c5dd15c2aa8b6231e1"} Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.290356 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-xxmsm" Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.297934 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.297913201 podStartE2EDuration="8.297913201s" podCreationTimestamp="2026-03-13 09:36:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:49.289606603 +0000 UTC m=+1450.039521280" watchObservedRunningTime="2026-03-13 09:36:49.297913201 +0000 UTC m=+1450.047827878" Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.337966 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=17.337948809 podStartE2EDuration="17.337948809s" podCreationTimestamp="2026-03-13 09:36:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:49.329155599 +0000 UTC m=+1450.079070276" watchObservedRunningTime="2026-03-13 09:36:49.337948809 +0000 UTC m=+1450.087863486" Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.413217 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=6.413182948 podStartE2EDuration="6.413182948s" podCreationTimestamp="2026-03-13 09:36:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:36:49.39187489 +0000 UTC m=+1450.141789577" watchObservedRunningTime="2026-03-13 09:36:49.413182948 +0000 UTC m=+1450.163097625" Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.851587 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.969767 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqksg\" (UniqueName: \"kubernetes.io/projected/4a15fab1-c914-4555-813a-ab9552b29a22-kube-api-access-dqksg\") pod \"4a15fab1-c914-4555-813a-ab9552b29a22\" (UID: \"4a15fab1-c914-4555-813a-ab9552b29a22\") " Mar 13 09:36:49 crc kubenswrapper[4930]: I0313 09:36:49.970036 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a15fab1-c914-4555-813a-ab9552b29a22-operator-scripts\") pod \"4a15fab1-c914-4555-813a-ab9552b29a22\" (UID: \"4a15fab1-c914-4555-813a-ab9552b29a22\") " Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.303408 4930 scope.go:117] "RemoveContainer" containerID="ef9d3b908088bcbe51f586c23e75287474186c5ddf9f52b7ab287e61a55a9477" Mar 13 09:36:50 crc kubenswrapper[4930]: E0313 09:36:50.303966 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-66d7f6c554-7d4pc_openstack(17c2545f-f1f1-4c0a-a689-73202f48ddf6)\"" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" podUID="17c2545f-f1f1-4c0a-a689-73202f48ddf6" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.304519 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" event={"ID":"4a15fab1-c914-4555-813a-ab9552b29a22","Type":"ContainerDied","Data":"98653f43710ccbbef2f16818243db0fd693465f968e3d9bd3026ef6ae078899d"} Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.304545 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98653f43710ccbbef2f16818243db0fd693465f968e3d9bd3026ef6ae078899d" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.304591 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-7e6b-account-create-update-tgmkk" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.308729 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e5091765-b1c8-43be-9579-d503faf84b6e","Type":"ContainerStarted","Data":"986b3303472f21fe517b74cfcb79be8696633be646718bf66fa2a15136e71294"} Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.309590 4930 scope.go:117] "RemoveContainer" containerID="cdb8e49793e92f7f6013638a2391f36b891181793818f70e69c8578f03c9901c" Mar 13 09:36:50 crc kubenswrapper[4930]: E0313 09:36:50.309894 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7889c489fb-hj2f9_openstack(304ba316-73ff-420c-af72-945f29aa5e2e)\"" pod="openstack/heat-api-7889c489fb-hj2f9" podUID="304ba316-73ff-420c-af72-945f29aa5e2e" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.504613 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a15fab1-c914-4555-813a-ab9552b29a22-kube-api-access-dqksg" (OuterVolumeSpecName: "kube-api-access-dqksg") pod "4a15fab1-c914-4555-813a-ab9552b29a22" (UID: "4a15fab1-c914-4555-813a-ab9552b29a22"). InnerVolumeSpecName "kube-api-access-dqksg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.511269 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a15fab1-c914-4555-813a-ab9552b29a22-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4a15fab1-c914-4555-813a-ab9552b29a22" (UID: "4a15fab1-c914-4555-813a-ab9552b29a22"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.586330 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a15fab1-c914-4555-813a-ab9552b29a22-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.586508 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqksg\" (UniqueName: \"kubernetes.io/projected/4a15fab1-c914-4555-813a-ab9552b29a22-kube-api-access-dqksg\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.663794 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lzrnq" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.668919 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b5f3-account-create-update-9zsnn" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.687799 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9e12cb85-3947-4c60-96a6-8ac07f5e63e8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9e12cb85-3947-4c60-96a6-8ac07f5e63e8" (UID: "9e12cb85-3947-4c60-96a6-8ac07f5e63e8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.687850 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e12cb85-3947-4c60-96a6-8ac07f5e63e8-operator-scripts\") pod \"9e12cb85-3947-4c60-96a6-8ac07f5e63e8\" (UID: \"9e12cb85-3947-4c60-96a6-8ac07f5e63e8\") " Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.688051 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gq2ll\" (UniqueName: \"kubernetes.io/projected/9e12cb85-3947-4c60-96a6-8ac07f5e63e8-kube-api-access-gq2ll\") pod \"9e12cb85-3947-4c60-96a6-8ac07f5e63e8\" (UID: \"9e12cb85-3947-4c60-96a6-8ac07f5e63e8\") " Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.689314 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9e12cb85-3947-4c60-96a6-8ac07f5e63e8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.700152 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e12cb85-3947-4c60-96a6-8ac07f5e63e8-kube-api-access-gq2ll" (OuterVolumeSpecName: "kube-api-access-gq2ll") pod "9e12cb85-3947-4c60-96a6-8ac07f5e63e8" (UID: "9e12cb85-3947-4c60-96a6-8ac07f5e63e8"). InnerVolumeSpecName "kube-api-access-gq2ll". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.791418 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a92113b-21a5-4db3-9708-589fc0aa57e8-operator-scripts\") pod \"4a92113b-21a5-4db3-9708-589fc0aa57e8\" (UID: \"4a92113b-21a5-4db3-9708-589fc0aa57e8\") " Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.791633 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h84lg\" (UniqueName: \"kubernetes.io/projected/4a92113b-21a5-4db3-9708-589fc0aa57e8-kube-api-access-h84lg\") pod \"4a92113b-21a5-4db3-9708-589fc0aa57e8\" (UID: \"4a92113b-21a5-4db3-9708-589fc0aa57e8\") " Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.792057 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a92113b-21a5-4db3-9708-589fc0aa57e8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4a92113b-21a5-4db3-9708-589fc0aa57e8" (UID: "4a92113b-21a5-4db3-9708-589fc0aa57e8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.792818 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gq2ll\" (UniqueName: \"kubernetes.io/projected/9e12cb85-3947-4c60-96a6-8ac07f5e63e8-kube-api-access-gq2ll\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.792845 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4a92113b-21a5-4db3-9708-589fc0aa57e8-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.796593 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a92113b-21a5-4db3-9708-589fc0aa57e8-kube-api-access-h84lg" (OuterVolumeSpecName: "kube-api-access-h84lg") pod "4a92113b-21a5-4db3-9708-589fc0aa57e8" (UID: "4a92113b-21a5-4db3-9708-589fc0aa57e8"). InnerVolumeSpecName "kube-api-access-h84lg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:50 crc kubenswrapper[4930]: I0313 09:36:50.894653 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h84lg\" (UniqueName: \"kubernetes.io/projected/4a92113b-21a5-4db3-9708-589fc0aa57e8-kube-api-access-h84lg\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:51 crc kubenswrapper[4930]: I0313 09:36:51.322798 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-lzrnq" event={"ID":"9e12cb85-3947-4c60-96a6-8ac07f5e63e8","Type":"ContainerDied","Data":"3b38c16cad245857ce3b91e4f3cb3027749ec8a011d0d5ea30a2e543f318b8db"} Mar 13 09:36:51 crc kubenswrapper[4930]: I0313 09:36:51.323183 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3b38c16cad245857ce3b91e4f3cb3027749ec8a011d0d5ea30a2e543f318b8db" Mar 13 09:36:51 crc kubenswrapper[4930]: I0313 09:36:51.322806 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-lzrnq" Mar 13 09:36:51 crc kubenswrapper[4930]: I0313 09:36:51.330607 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-b5f3-account-create-update-9zsnn" Mar 13 09:36:51 crc kubenswrapper[4930]: I0313 09:36:51.330658 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-b5f3-account-create-update-9zsnn" event={"ID":"4a92113b-21a5-4db3-9708-589fc0aa57e8","Type":"ContainerDied","Data":"7fb40d52130773d34237fbfe656f5ada4da1abeabb2c2598579ce8bcf4e5c164"} Mar 13 09:36:51 crc kubenswrapper[4930]: I0313 09:36:51.330695 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7fb40d52130773d34237fbfe656f5ada4da1abeabb2c2598579ce8bcf4e5c164" Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.161605 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.241297 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-7889c489fb-hj2f9"] Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.370407 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="ceilometer-central-agent" containerID="cri-o://0c5ac89688d9ba42713fc6bb044f42a88503026717ee271ac665ad3740a28374" gracePeriod=30 Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.370825 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="proxy-httpd" containerID="cri-o://c72579b768d4f083d859a30f12ea82ceb70d718b9481d0549da6f86d0fd0eb30" gracePeriod=30 Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.370874 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="ceilometer-notification-agent" containerID="cri-o://8f167a7939d155e62597044e25b13583804dfa7aa665d5097cfc461951893973" gracePeriod=30 Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.370924 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="sg-core" containerID="cri-o://986b3303472f21fe517b74cfcb79be8696633be646718bf66fa2a15136e71294" gracePeriod=30 Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.370634 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e5091765-b1c8-43be-9579-d503faf84b6e","Type":"ContainerStarted","Data":"c72579b768d4f083d859a30f12ea82ceb70d718b9481d0549da6f86d0fd0eb30"} Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.371662 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.408525 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=12.761944975 podStartE2EDuration="20.408505445s" podCreationTimestamp="2026-03-13 09:36:32 +0000 UTC" firstStartedPulling="2026-03-13 09:36:43.56294916 +0000 UTC m=+1444.312863837" lastFinishedPulling="2026-03-13 09:36:51.20950963 +0000 UTC m=+1451.959424307" observedRunningTime="2026-03-13 09:36:52.406734999 +0000 UTC m=+1453.156649676" watchObservedRunningTime="2026-03-13 09:36:52.408505445 +0000 UTC m=+1453.158420122" Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.426317 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.427618 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.504825 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.506925 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.903971 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.947523 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdhkg\" (UniqueName: \"kubernetes.io/projected/304ba316-73ff-420c-af72-945f29aa5e2e-kube-api-access-zdhkg\") pod \"304ba316-73ff-420c-af72-945f29aa5e2e\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.947590 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-config-data\") pod \"304ba316-73ff-420c-af72-945f29aa5e2e\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.947781 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-combined-ca-bundle\") pod \"304ba316-73ff-420c-af72-945f29aa5e2e\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.947819 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-config-data-custom\") pod \"304ba316-73ff-420c-af72-945f29aa5e2e\" (UID: \"304ba316-73ff-420c-af72-945f29aa5e2e\") " Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.952856 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/304ba316-73ff-420c-af72-945f29aa5e2e-kube-api-access-zdhkg" (OuterVolumeSpecName: "kube-api-access-zdhkg") pod "304ba316-73ff-420c-af72-945f29aa5e2e" (UID: "304ba316-73ff-420c-af72-945f29aa5e2e"). InnerVolumeSpecName "kube-api-access-zdhkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.953178 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "304ba316-73ff-420c-af72-945f29aa5e2e" (UID: "304ba316-73ff-420c-af72-945f29aa5e2e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:52 crc kubenswrapper[4930]: I0313 09:36:52.980210 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "304ba316-73ff-420c-af72-945f29aa5e2e" (UID: "304ba316-73ff-420c-af72-945f29aa5e2e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.009982 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-config-data" (OuterVolumeSpecName: "config-data") pod "304ba316-73ff-420c-af72-945f29aa5e2e" (UID: "304ba316-73ff-420c-af72-945f29aa5e2e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.050932 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdhkg\" (UniqueName: \"kubernetes.io/projected/304ba316-73ff-420c-af72-945f29aa5e2e-kube-api-access-zdhkg\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.050964 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.050974 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.050982 4930 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/304ba316-73ff-420c-af72-945f29aa5e2e-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.384715 4930 generic.go:334] "Generic (PLEG): container finished" podID="e5091765-b1c8-43be-9579-d503faf84b6e" containerID="c72579b768d4f083d859a30f12ea82ceb70d718b9481d0549da6f86d0fd0eb30" exitCode=0 Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.385084 4930 generic.go:334] "Generic (PLEG): container finished" podID="e5091765-b1c8-43be-9579-d503faf84b6e" containerID="986b3303472f21fe517b74cfcb79be8696633be646718bf66fa2a15136e71294" exitCode=2 Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.385096 4930 generic.go:334] "Generic (PLEG): container finished" podID="e5091765-b1c8-43be-9579-d503faf84b6e" containerID="8f167a7939d155e62597044e25b13583804dfa7aa665d5097cfc461951893973" exitCode=0 Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.385106 4930 generic.go:334] "Generic (PLEG): container finished" podID="e5091765-b1c8-43be-9579-d503faf84b6e" containerID="0c5ac89688d9ba42713fc6bb044f42a88503026717ee271ac665ad3740a28374" exitCode=0 Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.384889 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e5091765-b1c8-43be-9579-d503faf84b6e","Type":"ContainerDied","Data":"c72579b768d4f083d859a30f12ea82ceb70d718b9481d0549da6f86d0fd0eb30"} Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.385176 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e5091765-b1c8-43be-9579-d503faf84b6e","Type":"ContainerDied","Data":"986b3303472f21fe517b74cfcb79be8696633be646718bf66fa2a15136e71294"} Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.385191 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e5091765-b1c8-43be-9579-d503faf84b6e","Type":"ContainerDied","Data":"8f167a7939d155e62597044e25b13583804dfa7aa665d5097cfc461951893973"} Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.385203 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e5091765-b1c8-43be-9579-d503faf84b6e","Type":"ContainerDied","Data":"0c5ac89688d9ba42713fc6bb044f42a88503026717ee271ac665ad3740a28374"} Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.401817 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7889c489fb-hj2f9" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.403429 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7889c489fb-hj2f9" event={"ID":"304ba316-73ff-420c-af72-945f29aa5e2e","Type":"ContainerDied","Data":"fc6e2d22a8fceecf478d19c8a447a42ce80ddf3c1589ffb1bf9099ae6abbe3bd"} Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.403486 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.403504 4930 scope.go:117] "RemoveContainer" containerID="cdb8e49793e92f7f6013638a2391f36b891181793818f70e69c8578f03c9901c" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.403804 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.456691 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-7889c489fb-hj2f9"] Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.481195 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-7889c489fb-hj2f9"] Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.510986 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.511020 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.511867 4930 scope.go:117] "RemoveContainer" containerID="ef9d3b908088bcbe51f586c23e75287474186c5ddf9f52b7ab287e61a55a9477" Mar 13 09:36:53 crc kubenswrapper[4930]: E0313 09:36:53.512148 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-66d7f6c554-7d4pc_openstack(17c2545f-f1f1-4c0a-a689-73202f48ddf6)\"" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" podUID="17c2545f-f1f1-4c0a-a689-73202f48ddf6" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.600211 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.600348 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.633719 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.646231 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.735053 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.866165 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-sg-core-conf-yaml\") pod \"e5091765-b1c8-43be-9579-d503faf84b6e\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.866256 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-scripts\") pod \"e5091765-b1c8-43be-9579-d503faf84b6e\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.866299 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-config-data\") pod \"e5091765-b1c8-43be-9579-d503faf84b6e\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.866331 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xmh5s\" (UniqueName: \"kubernetes.io/projected/e5091765-b1c8-43be-9579-d503faf84b6e-kube-api-access-xmh5s\") pod \"e5091765-b1c8-43be-9579-d503faf84b6e\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.866462 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-combined-ca-bundle\") pod \"e5091765-b1c8-43be-9579-d503faf84b6e\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.866492 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e5091765-b1c8-43be-9579-d503faf84b6e-log-httpd\") pod \"e5091765-b1c8-43be-9579-d503faf84b6e\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.866635 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e5091765-b1c8-43be-9579-d503faf84b6e-run-httpd\") pod \"e5091765-b1c8-43be-9579-d503faf84b6e\" (UID: \"e5091765-b1c8-43be-9579-d503faf84b6e\") " Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.867508 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5091765-b1c8-43be-9579-d503faf84b6e-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e5091765-b1c8-43be-9579-d503faf84b6e" (UID: "e5091765-b1c8-43be-9579-d503faf84b6e"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.874557 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5091765-b1c8-43be-9579-d503faf84b6e-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e5091765-b1c8-43be-9579-d503faf84b6e" (UID: "e5091765-b1c8-43be-9579-d503faf84b6e"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.885678 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-scripts" (OuterVolumeSpecName: "scripts") pod "e5091765-b1c8-43be-9579-d503faf84b6e" (UID: "e5091765-b1c8-43be-9579-d503faf84b6e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.915045 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5091765-b1c8-43be-9579-d503faf84b6e-kube-api-access-xmh5s" (OuterVolumeSpecName: "kube-api-access-xmh5s") pod "e5091765-b1c8-43be-9579-d503faf84b6e" (UID: "e5091765-b1c8-43be-9579-d503faf84b6e"). InnerVolumeSpecName "kube-api-access-xmh5s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.969103 4930 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e5091765-b1c8-43be-9579-d503faf84b6e-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.969143 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.969153 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xmh5s\" (UniqueName: \"kubernetes.io/projected/e5091765-b1c8-43be-9579-d503faf84b6e-kube-api-access-xmh5s\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.969162 4930 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e5091765-b1c8-43be-9579-d503faf84b6e-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:53 crc kubenswrapper[4930]: I0313 09:36:53.998912 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="304ba316-73ff-420c-af72-945f29aa5e2e" path="/var/lib/kubelet/pods/304ba316-73ff-420c-af72-945f29aa5e2e/volumes" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.001603 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e5091765-b1c8-43be-9579-d503faf84b6e" (UID: "e5091765-b1c8-43be-9579-d503faf84b6e"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.071963 4930 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.079738 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5091765-b1c8-43be-9579-d503faf84b6e" (UID: "e5091765-b1c8-43be-9579-d503faf84b6e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.083731 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-config-data" (OuterVolumeSpecName: "config-data") pod "e5091765-b1c8-43be-9579-d503faf84b6e" (UID: "e5091765-b1c8-43be-9579-d503faf84b6e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.173894 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.173940 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5091765-b1c8-43be-9579-d503faf84b6e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.308165 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gb9wl"] Mar 13 09:36:54 crc kubenswrapper[4930]: E0313 09:36:54.308846 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="304ba316-73ff-420c-af72-945f29aa5e2e" containerName="heat-api" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.308860 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="304ba316-73ff-420c-af72-945f29aa5e2e" containerName="heat-api" Mar 13 09:36:54 crc kubenswrapper[4930]: E0313 09:36:54.308884 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a15fab1-c914-4555-813a-ab9552b29a22" containerName="mariadb-account-create-update" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.308890 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a15fab1-c914-4555-813a-ab9552b29a22" containerName="mariadb-account-create-update" Mar 13 09:36:54 crc kubenswrapper[4930]: E0313 09:36:54.308917 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ec86dec2-eb5a-497e-bba2-d1b96d80c65c" containerName="mariadb-database-create" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.308924 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ec86dec2-eb5a-497e-bba2-d1b96d80c65c" containerName="mariadb-database-create" Mar 13 09:36:54 crc kubenswrapper[4930]: E0313 09:36:54.308930 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="ceilometer-notification-agent" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.308936 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="ceilometer-notification-agent" Mar 13 09:36:54 crc kubenswrapper[4930]: E0313 09:36:54.308947 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3c2e237-7f4c-48a2-8a8d-f5838192decc" containerName="mariadb-database-create" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.308953 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3c2e237-7f4c-48a2-8a8d-f5838192decc" containerName="mariadb-database-create" Mar 13 09:36:54 crc kubenswrapper[4930]: E0313 09:36:54.308962 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e12cb85-3947-4c60-96a6-8ac07f5e63e8" containerName="mariadb-database-create" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.308970 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e12cb85-3947-4c60-96a6-8ac07f5e63e8" containerName="mariadb-database-create" Mar 13 09:36:54 crc kubenswrapper[4930]: E0313 09:36:54.308980 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5942b4a-ef6d-40ca-b67f-e7965513004d" containerName="mariadb-account-create-update" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.308986 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5942b4a-ef6d-40ca-b67f-e7965513004d" containerName="mariadb-account-create-update" Mar 13 09:36:54 crc kubenswrapper[4930]: E0313 09:36:54.308993 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="sg-core" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.308998 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="sg-core" Mar 13 09:36:54 crc kubenswrapper[4930]: E0313 09:36:54.309007 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a92113b-21a5-4db3-9708-589fc0aa57e8" containerName="mariadb-account-create-update" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309012 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a92113b-21a5-4db3-9708-589fc0aa57e8" containerName="mariadb-account-create-update" Mar 13 09:36:54 crc kubenswrapper[4930]: E0313 09:36:54.309021 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="proxy-httpd" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309026 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="proxy-httpd" Mar 13 09:36:54 crc kubenswrapper[4930]: E0313 09:36:54.309039 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="ceilometer-central-agent" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309045 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="ceilometer-central-agent" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309264 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a92113b-21a5-4db3-9708-589fc0aa57e8" containerName="mariadb-account-create-update" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309276 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e12cb85-3947-4c60-96a6-8ac07f5e63e8" containerName="mariadb-database-create" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309292 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3c2e237-7f4c-48a2-8a8d-f5838192decc" containerName="mariadb-database-create" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309306 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="ceilometer-central-agent" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309315 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="sg-core" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309322 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="304ba316-73ff-420c-af72-945f29aa5e2e" containerName="heat-api" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309332 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5942b4a-ef6d-40ca-b67f-e7965513004d" containerName="mariadb-account-create-update" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309343 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ec86dec2-eb5a-497e-bba2-d1b96d80c65c" containerName="mariadb-database-create" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309355 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="ceilometer-notification-agent" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309363 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" containerName="proxy-httpd" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309376 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="304ba316-73ff-420c-af72-945f29aa5e2e" containerName="heat-api" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.309386 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a15fab1-c914-4555-813a-ab9552b29a22" containerName="mariadb-account-create-update" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.310280 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.312708 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pdj49" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.314903 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.320812 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gb9wl"] Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.322423 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.418080 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e5091765-b1c8-43be-9579-d503faf84b6e","Type":"ContainerDied","Data":"c3b574ad5214038a41c7a4cbd2590943f742ff6b249f2702947402114174d225"} Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.418131 4930 scope.go:117] "RemoveContainer" containerID="c72579b768d4f083d859a30f12ea82ceb70d718b9481d0549da6f86d0fd0eb30" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.418278 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.427766 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.428083 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.457683 4930 scope.go:117] "RemoveContainer" containerID="986b3303472f21fe517b74cfcb79be8696633be646718bf66fa2a15136e71294" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.465319 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.486674 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gb9wl\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.486795 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-scripts\") pod \"nova-cell0-conductor-db-sync-gb9wl\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.486874 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-config-data\") pod \"nova-cell0-conductor-db-sync-gb9wl\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.487103 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzp98\" (UniqueName: \"kubernetes.io/projected/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-kube-api-access-xzp98\") pod \"nova-cell0-conductor-db-sync-gb9wl\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.487713 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.496416 4930 scope.go:117] "RemoveContainer" containerID="8f167a7939d155e62597044e25b13583804dfa7aa665d5097cfc461951893973" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.504074 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:54 crc kubenswrapper[4930]: E0313 09:36:54.504756 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="304ba316-73ff-420c-af72-945f29aa5e2e" containerName="heat-api" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.504777 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="304ba316-73ff-420c-af72-945f29aa5e2e" containerName="heat-api" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.511143 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.519813 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.520068 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.522159 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.589297 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.589369 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-scripts\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.589475 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzp98\" (UniqueName: \"kubernetes.io/projected/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-kube-api-access-xzp98\") pod \"nova-cell0-conductor-db-sync-gb9wl\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.589557 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gb9wl\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.589594 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74mxz\" (UniqueName: \"kubernetes.io/projected/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-kube-api-access-74mxz\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.589650 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-run-httpd\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.589667 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-scripts\") pod \"nova-cell0-conductor-db-sync-gb9wl\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.589703 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-config-data\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.589735 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-config-data\") pod \"nova-cell0-conductor-db-sync-gb9wl\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.589809 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.589835 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-log-httpd\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.594063 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-gb9wl\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.596028 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-config-data\") pod \"nova-cell0-conductor-db-sync-gb9wl\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.599624 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-scripts\") pod \"nova-cell0-conductor-db-sync-gb9wl\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.602954 4930 scope.go:117] "RemoveContainer" containerID="0c5ac89688d9ba42713fc6bb044f42a88503026717ee271ac665ad3740a28374" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.614894 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzp98\" (UniqueName: \"kubernetes.io/projected/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-kube-api-access-xzp98\") pod \"nova-cell0-conductor-db-sync-gb9wl\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.647495 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.694459 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.694543 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-scripts\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.694726 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74mxz\" (UniqueName: \"kubernetes.io/projected/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-kube-api-access-74mxz\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.694818 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-run-httpd\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.694876 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-config-data\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.695028 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.695072 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-log-httpd\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.695680 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-log-httpd\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.695990 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-run-httpd\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.700282 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.703641 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.710531 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.711231 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-scripts\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.715858 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-config-data\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.721157 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74mxz\" (UniqueName: \"kubernetes.io/projected/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-kube-api-access-74mxz\") pod \"ceilometer-0\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " pod="openstack/ceilometer-0" Mar 13 09:36:54 crc kubenswrapper[4930]: I0313 09:36:54.853172 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:36:55 crc kubenswrapper[4930]: I0313 09:36:55.186266 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gb9wl"] Mar 13 09:36:55 crc kubenswrapper[4930]: I0313 09:36:55.442364 4930 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 09:36:55 crc kubenswrapper[4930]: I0313 09:36:55.442392 4930 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 09:36:55 crc kubenswrapper[4930]: I0313 09:36:55.443392 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gb9wl" event={"ID":"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0","Type":"ContainerStarted","Data":"158e96e9538bce86be4b069370b744406c6584a19475b586d34622caa5f0aa64"} Mar 13 09:36:55 crc kubenswrapper[4930]: I0313 09:36:55.486301 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:36:56 crc kubenswrapper[4930]: I0313 09:36:56.009994 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5091765-b1c8-43be-9579-d503faf84b6e" path="/var/lib/kubelet/pods/e5091765-b1c8-43be-9579-d503faf84b6e/volumes" Mar 13 09:36:56 crc kubenswrapper[4930]: I0313 09:36:56.368848 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:36:56 crc kubenswrapper[4930]: I0313 09:36:56.433867 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:36:56 crc kubenswrapper[4930]: I0313 09:36:56.461963 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1","Type":"ContainerStarted","Data":"f3ea9efefa1ee8bc77fede0c1ab621eb62ef627679be39ac6be050488b91c9d1"} Mar 13 09:36:56 crc kubenswrapper[4930]: I0313 09:36:56.461998 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1","Type":"ContainerStarted","Data":"1e41b87ceb436310e4b962cd08b636c07b76431910a266d1cd14e9bc963ae10f"} Mar 13 09:36:56 crc kubenswrapper[4930]: I0313 09:36:56.613141 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bsl74"] Mar 13 09:36:57 crc kubenswrapper[4930]: I0313 09:36:57.488567 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bsl74" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerName="registry-server" containerID="cri-o://1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9" gracePeriod=2 Mar 13 09:36:57 crc kubenswrapper[4930]: I0313 09:36:57.489362 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1","Type":"ContainerStarted","Data":"22eb7e9fb6ed9e0760dea6d091064a7f9c8ae2cb1cc36e6d44cff8930f973265"} Mar 13 09:36:57 crc kubenswrapper[4930]: I0313 09:36:57.556010 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:36:57 crc kubenswrapper[4930]: I0313 09:36:57.640612 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-66d7f6c554-7d4pc"] Mar 13 09:36:57 crc kubenswrapper[4930]: I0313 09:36:57.845995 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 09:36:57 crc kubenswrapper[4930]: I0313 09:36:57.846424 4930 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 09:36:57 crc kubenswrapper[4930]: I0313 09:36:57.909987 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:57 crc kubenswrapper[4930]: I0313 09:36:57.910125 4930 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 13 09:36:57 crc kubenswrapper[4930]: I0313 09:36:57.910312 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.054507 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.377817 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.390782 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.502548 4930 generic.go:334] "Generic (PLEG): container finished" podID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerID="1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9" exitCode=0 Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.502594 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bsl74" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.502617 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bsl74" event={"ID":"a73c1bd6-dd06-4e14-b8ee-1916942e53f5","Type":"ContainerDied","Data":"1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9"} Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.502644 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bsl74" event={"ID":"a73c1bd6-dd06-4e14-b8ee-1916942e53f5","Type":"ContainerDied","Data":"0352d4207fd23d7f82439fb2d1db720a090328debdaa9c88d25c5404d09edf7f"} Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.502661 4930 scope.go:117] "RemoveContainer" containerID="1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.508374 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1","Type":"ContainerStarted","Data":"47e2399fe2f9c2c121e91b37207267e5116a732e38613aceaab6b815ea517b3e"} Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.510847 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.511282 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-66d7f6c554-7d4pc" event={"ID":"17c2545f-f1f1-4c0a-a689-73202f48ddf6","Type":"ContainerDied","Data":"5086aacf2d28d8b560aa097b6a53ab3b7d62cb536a3970c14d6255d4055fc66d"} Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.518389 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-utilities\") pod \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\" (UID: \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\") " Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.518497 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9n6vz\" (UniqueName: \"kubernetes.io/projected/17c2545f-f1f1-4c0a-a689-73202f48ddf6-kube-api-access-9n6vz\") pod \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.518543 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svgvd\" (UniqueName: \"kubernetes.io/projected/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-kube-api-access-svgvd\") pod \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\" (UID: \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\") " Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.519150 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-utilities" (OuterVolumeSpecName: "utilities") pod "a73c1bd6-dd06-4e14-b8ee-1916942e53f5" (UID: "a73c1bd6-dd06-4e14-b8ee-1916942e53f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.519469 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-catalog-content\") pod \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\" (UID: \"a73c1bd6-dd06-4e14-b8ee-1916942e53f5\") " Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.519565 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-combined-ca-bundle\") pod \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.519688 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-config-data-custom\") pod \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.519785 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-config-data\") pod \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\" (UID: \"17c2545f-f1f1-4c0a-a689-73202f48ddf6\") " Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.520342 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.525442 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "17c2545f-f1f1-4c0a-a689-73202f48ddf6" (UID: "17c2545f-f1f1-4c0a-a689-73202f48ddf6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.525602 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17c2545f-f1f1-4c0a-a689-73202f48ddf6-kube-api-access-9n6vz" (OuterVolumeSpecName: "kube-api-access-9n6vz") pod "17c2545f-f1f1-4c0a-a689-73202f48ddf6" (UID: "17c2545f-f1f1-4c0a-a689-73202f48ddf6"). InnerVolumeSpecName "kube-api-access-9n6vz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.528644 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-kube-api-access-svgvd" (OuterVolumeSpecName: "kube-api-access-svgvd") pod "a73c1bd6-dd06-4e14-b8ee-1916942e53f5" (UID: "a73c1bd6-dd06-4e14-b8ee-1916942e53f5"). InnerVolumeSpecName "kube-api-access-svgvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.548016 4930 scope.go:117] "RemoveContainer" containerID="3f4bcb33f157da3b03c1334f17a6c246e3568f1a75cd98ca29fce9755dba763c" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.589527 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17c2545f-f1f1-4c0a-a689-73202f48ddf6" (UID: "17c2545f-f1f1-4c0a-a689-73202f48ddf6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.592647 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-config-data" (OuterVolumeSpecName: "config-data") pod "17c2545f-f1f1-4c0a-a689-73202f48ddf6" (UID: "17c2545f-f1f1-4c0a-a689-73202f48ddf6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.625627 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9n6vz\" (UniqueName: \"kubernetes.io/projected/17c2545f-f1f1-4c0a-a689-73202f48ddf6-kube-api-access-9n6vz\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.625663 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svgvd\" (UniqueName: \"kubernetes.io/projected/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-kube-api-access-svgvd\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.625674 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.625683 4930 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.625692 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17c2545f-f1f1-4c0a-a689-73202f48ddf6-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.651302 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a73c1bd6-dd06-4e14-b8ee-1916942e53f5" (UID: "a73c1bd6-dd06-4e14-b8ee-1916942e53f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.683727 4930 scope.go:117] "RemoveContainer" containerID="96782db3f1a05ce85427f059e78463d435d238cf3a17f8dfe64c138a48f73674" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.709426 4930 scope.go:117] "RemoveContainer" containerID="1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9" Mar 13 09:36:58 crc kubenswrapper[4930]: E0313 09:36:58.712026 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9\": container with ID starting with 1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9 not found: ID does not exist" containerID="1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.712086 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9"} err="failed to get container status \"1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9\": rpc error: code = NotFound desc = could not find container \"1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9\": container with ID starting with 1305deeaa1ec7167790132354bdd69f53dc4f2b809ce38599adc5e75591d27e9 not found: ID does not exist" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.712120 4930 scope.go:117] "RemoveContainer" containerID="3f4bcb33f157da3b03c1334f17a6c246e3568f1a75cd98ca29fce9755dba763c" Mar 13 09:36:58 crc kubenswrapper[4930]: E0313 09:36:58.712625 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f4bcb33f157da3b03c1334f17a6c246e3568f1a75cd98ca29fce9755dba763c\": container with ID starting with 3f4bcb33f157da3b03c1334f17a6c246e3568f1a75cd98ca29fce9755dba763c not found: ID does not exist" containerID="3f4bcb33f157da3b03c1334f17a6c246e3568f1a75cd98ca29fce9755dba763c" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.712654 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f4bcb33f157da3b03c1334f17a6c246e3568f1a75cd98ca29fce9755dba763c"} err="failed to get container status \"3f4bcb33f157da3b03c1334f17a6c246e3568f1a75cd98ca29fce9755dba763c\": rpc error: code = NotFound desc = could not find container \"3f4bcb33f157da3b03c1334f17a6c246e3568f1a75cd98ca29fce9755dba763c\": container with ID starting with 3f4bcb33f157da3b03c1334f17a6c246e3568f1a75cd98ca29fce9755dba763c not found: ID does not exist" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.712671 4930 scope.go:117] "RemoveContainer" containerID="96782db3f1a05ce85427f059e78463d435d238cf3a17f8dfe64c138a48f73674" Mar 13 09:36:58 crc kubenswrapper[4930]: E0313 09:36:58.712972 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96782db3f1a05ce85427f059e78463d435d238cf3a17f8dfe64c138a48f73674\": container with ID starting with 96782db3f1a05ce85427f059e78463d435d238cf3a17f8dfe64c138a48f73674 not found: ID does not exist" containerID="96782db3f1a05ce85427f059e78463d435d238cf3a17f8dfe64c138a48f73674" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.712999 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96782db3f1a05ce85427f059e78463d435d238cf3a17f8dfe64c138a48f73674"} err="failed to get container status \"96782db3f1a05ce85427f059e78463d435d238cf3a17f8dfe64c138a48f73674\": rpc error: code = NotFound desc = could not find container \"96782db3f1a05ce85427f059e78463d435d238cf3a17f8dfe64c138a48f73674\": container with ID starting with 96782db3f1a05ce85427f059e78463d435d238cf3a17f8dfe64c138a48f73674 not found: ID does not exist" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.713018 4930 scope.go:117] "RemoveContainer" containerID="ef9d3b908088bcbe51f586c23e75287474186c5ddf9f52b7ab287e61a55a9477" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.727397 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a73c1bd6-dd06-4e14-b8ee-1916942e53f5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.869826 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bsl74"] Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.888508 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bsl74"] Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.904756 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-66d7f6c554-7d4pc"] Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.919960 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-66d7f6c554-7d4pc"] Mar 13 09:36:58 crc kubenswrapper[4930]: I0313 09:36:58.966596 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="d82ac246-17a4-4f81-acce-70aaa7d8f350" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.228:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 09:37:00 crc kubenswrapper[4930]: I0313 09:37:00.003832 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17c2545f-f1f1-4c0a-a689-73202f48ddf6" path="/var/lib/kubelet/pods/17c2545f-f1f1-4c0a-a689-73202f48ddf6/volumes" Mar 13 09:37:00 crc kubenswrapper[4930]: I0313 09:37:00.005156 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" path="/var/lib/kubelet/pods/a73c1bd6-dd06-4e14-b8ee-1916942e53f5/volumes" Mar 13 09:37:00 crc kubenswrapper[4930]: I0313 09:37:00.550085 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1","Type":"ContainerStarted","Data":"7f1a81491eb2ffdca73ab1e2e07dff21468f4489575e6ecff44520e3e6c7bf34"} Mar 13 09:37:00 crc kubenswrapper[4930]: I0313 09:37:00.550243 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="ceilometer-central-agent" containerID="cri-o://f3ea9efefa1ee8bc77fede0c1ab621eb62ef627679be39ac6be050488b91c9d1" gracePeriod=30 Mar 13 09:37:00 crc kubenswrapper[4930]: I0313 09:37:00.550371 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="proxy-httpd" containerID="cri-o://7f1a81491eb2ffdca73ab1e2e07dff21468f4489575e6ecff44520e3e6c7bf34" gracePeriod=30 Mar 13 09:37:00 crc kubenswrapper[4930]: I0313 09:37:00.550418 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="sg-core" containerID="cri-o://47e2399fe2f9c2c121e91b37207267e5116a732e38613aceaab6b815ea517b3e" gracePeriod=30 Mar 13 09:37:00 crc kubenswrapper[4930]: I0313 09:37:00.550479 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="ceilometer-notification-agent" containerID="cri-o://22eb7e9fb6ed9e0760dea6d091064a7f9c8ae2cb1cc36e6d44cff8930f973265" gracePeriod=30 Mar 13 09:37:00 crc kubenswrapper[4930]: I0313 09:37:00.550752 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 09:37:00 crc kubenswrapper[4930]: I0313 09:37:00.582211 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.774232385 podStartE2EDuration="6.582188775s" podCreationTimestamp="2026-03-13 09:36:54 +0000 UTC" firstStartedPulling="2026-03-13 09:36:55.487142544 +0000 UTC m=+1456.237057221" lastFinishedPulling="2026-03-13 09:36:59.295098934 +0000 UTC m=+1460.045013611" observedRunningTime="2026-03-13 09:37:00.574811952 +0000 UTC m=+1461.324726629" watchObservedRunningTime="2026-03-13 09:37:00.582188775 +0000 UTC m=+1461.332103452" Mar 13 09:37:00 crc kubenswrapper[4930]: I0313 09:37:00.985807 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 13 09:37:01 crc kubenswrapper[4930]: I0313 09:37:01.562475 4930 generic.go:334] "Generic (PLEG): container finished" podID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerID="7f1a81491eb2ffdca73ab1e2e07dff21468f4489575e6ecff44520e3e6c7bf34" exitCode=0 Mar 13 09:37:01 crc kubenswrapper[4930]: I0313 09:37:01.562505 4930 generic.go:334] "Generic (PLEG): container finished" podID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerID="47e2399fe2f9c2c121e91b37207267e5116a732e38613aceaab6b815ea517b3e" exitCode=2 Mar 13 09:37:01 crc kubenswrapper[4930]: I0313 09:37:01.562513 4930 generic.go:334] "Generic (PLEG): container finished" podID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerID="22eb7e9fb6ed9e0760dea6d091064a7f9c8ae2cb1cc36e6d44cff8930f973265" exitCode=0 Mar 13 09:37:01 crc kubenswrapper[4930]: I0313 09:37:01.562533 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1","Type":"ContainerDied","Data":"7f1a81491eb2ffdca73ab1e2e07dff21468f4489575e6ecff44520e3e6c7bf34"} Mar 13 09:37:01 crc kubenswrapper[4930]: I0313 09:37:01.562557 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1","Type":"ContainerDied","Data":"47e2399fe2f9c2c121e91b37207267e5116a732e38613aceaab6b815ea517b3e"} Mar 13 09:37:01 crc kubenswrapper[4930]: I0313 09:37:01.562567 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1","Type":"ContainerDied","Data":"22eb7e9fb6ed9e0760dea6d091064a7f9c8ae2cb1cc36e6d44cff8930f973265"} Mar 13 09:37:03 crc kubenswrapper[4930]: I0313 09:37:03.529905 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:37:03 crc kubenswrapper[4930]: I0313 09:37:03.634536 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7f4b8495c9-mqbjk"] Mar 13 09:37:03 crc kubenswrapper[4930]: I0313 09:37:03.634815 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-7f4b8495c9-mqbjk" podUID="7754b93c-5665-4f0d-8e26-d9910d1e90c4" containerName="heat-engine" containerID="cri-o://db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e" gracePeriod=60 Mar 13 09:37:04 crc kubenswrapper[4930]: E0313 09:37:04.249574 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 09:37:04 crc kubenswrapper[4930]: E0313 09:37:04.251785 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 09:37:04 crc kubenswrapper[4930]: E0313 09:37:04.253843 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 09:37:04 crc kubenswrapper[4930]: E0313 09:37:04.253881 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-7f4b8495c9-mqbjk" podUID="7754b93c-5665-4f0d-8e26-d9910d1e90c4" containerName="heat-engine" Mar 13 09:37:09 crc kubenswrapper[4930]: I0313 09:37:09.708025 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gb9wl" event={"ID":"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0","Type":"ContainerStarted","Data":"93396c58766bfe9d2b827cfca22cf86170091e1e938e713952dfa6b7ed7f9479"} Mar 13 09:37:09 crc kubenswrapper[4930]: I0313 09:37:09.730600 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-gb9wl" podStartSLOduration=2.053318635 podStartE2EDuration="15.730576198s" podCreationTimestamp="2026-03-13 09:36:54 +0000 UTC" firstStartedPulling="2026-03-13 09:36:55.189620316 +0000 UTC m=+1455.939534983" lastFinishedPulling="2026-03-13 09:37:08.866877869 +0000 UTC m=+1469.616792546" observedRunningTime="2026-03-13 09:37:09.721597883 +0000 UTC m=+1470.471512560" watchObservedRunningTime="2026-03-13 09:37:09.730576198 +0000 UTC m=+1470.480490875" Mar 13 09:37:12 crc kubenswrapper[4930]: I0313 09:37:12.754063 4930 generic.go:334] "Generic (PLEG): container finished" podID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerID="f3ea9efefa1ee8bc77fede0c1ab621eb62ef627679be39ac6be050488b91c9d1" exitCode=0 Mar 13 09:37:12 crc kubenswrapper[4930]: I0313 09:37:12.754490 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1","Type":"ContainerDied","Data":"f3ea9efefa1ee8bc77fede0c1ab621eb62ef627679be39ac6be050488b91c9d1"} Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.215410 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.325129 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-log-httpd\") pod \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.325476 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-scripts\") pod \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.325545 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-combined-ca-bundle\") pod \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.325703 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-run-httpd\") pod \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.325824 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74mxz\" (UniqueName: \"kubernetes.io/projected/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-kube-api-access-74mxz\") pod \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.325865 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-config-data\") pod \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.325903 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-sg-core-conf-yaml\") pod \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\" (UID: \"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1\") " Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.328105 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" (UID: "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.329070 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" (UID: "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.332366 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-kube-api-access-74mxz" (OuterVolumeSpecName: "kube-api-access-74mxz") pod "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" (UID: "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1"). InnerVolumeSpecName "kube-api-access-74mxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.345282 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-scripts" (OuterVolumeSpecName: "scripts") pod "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" (UID: "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.367630 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" (UID: "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.428276 4930 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.428524 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.428589 4930 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.428645 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74mxz\" (UniqueName: \"kubernetes.io/projected/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-kube-api-access-74mxz\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.428702 4930 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.433591 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" (UID: "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.453145 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-config-data" (OuterVolumeSpecName: "config-data") pod "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" (UID: "359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.531295 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.531326 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.771887 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1","Type":"ContainerDied","Data":"1e41b87ceb436310e4b962cd08b636c07b76431910a266d1cd14e9bc963ae10f"} Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.771934 4930 scope.go:117] "RemoveContainer" containerID="7f1a81491eb2ffdca73ab1e2e07dff21468f4489575e6ecff44520e3e6c7bf34" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.772085 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.806201 4930 scope.go:117] "RemoveContainer" containerID="47e2399fe2f9c2c121e91b37207267e5116a732e38613aceaab6b815ea517b3e" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.821283 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.837243 4930 scope.go:117] "RemoveContainer" containerID="22eb7e9fb6ed9e0760dea6d091064a7f9c8ae2cb1cc36e6d44cff8930f973265" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.842962 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.855590 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:13 crc kubenswrapper[4930]: E0313 09:37:13.856145 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="sg-core" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856162 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="sg-core" Mar 13 09:37:13 crc kubenswrapper[4930]: E0313 09:37:13.856184 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerName="extract-content" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856191 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerName="extract-content" Mar 13 09:37:13 crc kubenswrapper[4930]: E0313 09:37:13.856208 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="ceilometer-notification-agent" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856214 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="ceilometer-notification-agent" Mar 13 09:37:13 crc kubenswrapper[4930]: E0313 09:37:13.856231 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17c2545f-f1f1-4c0a-a689-73202f48ddf6" containerName="heat-cfnapi" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856237 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="17c2545f-f1f1-4c0a-a689-73202f48ddf6" containerName="heat-cfnapi" Mar 13 09:37:13 crc kubenswrapper[4930]: E0313 09:37:13.856248 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerName="extract-utilities" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856255 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerName="extract-utilities" Mar 13 09:37:13 crc kubenswrapper[4930]: E0313 09:37:13.856264 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="proxy-httpd" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856270 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="proxy-httpd" Mar 13 09:37:13 crc kubenswrapper[4930]: E0313 09:37:13.856281 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="ceilometer-central-agent" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856287 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="ceilometer-central-agent" Mar 13 09:37:13 crc kubenswrapper[4930]: E0313 09:37:13.856300 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerName="registry-server" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856306 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerName="registry-server" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856547 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="ceilometer-notification-agent" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856587 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="ceilometer-central-agent" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856599 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="17c2545f-f1f1-4c0a-a689-73202f48ddf6" containerName="heat-cfnapi" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856610 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="17c2545f-f1f1-4c0a-a689-73202f48ddf6" containerName="heat-cfnapi" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856626 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="proxy-httpd" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856637 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="a73c1bd6-dd06-4e14-b8ee-1916942e53f5" containerName="registry-server" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856648 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" containerName="sg-core" Mar 13 09:37:13 crc kubenswrapper[4930]: E0313 09:37:13.856847 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17c2545f-f1f1-4c0a-a689-73202f48ddf6" containerName="heat-cfnapi" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.856855 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="17c2545f-f1f1-4c0a-a689-73202f48ddf6" containerName="heat-cfnapi" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.858821 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.863791 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.864018 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.871447 4930 scope.go:117] "RemoveContainer" containerID="f3ea9efefa1ee8bc77fede0c1ab621eb62ef627679be39ac6be050488b91c9d1" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.872515 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.939262 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.939345 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-run-httpd\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.939473 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-config-data\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.939504 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7tsm4\" (UniqueName: \"kubernetes.io/projected/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-kube-api-access-7tsm4\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.939538 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.939578 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-scripts\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.939689 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-log-httpd\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:13 crc kubenswrapper[4930]: I0313 09:37:13.982625 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1" path="/var/lib/kubelet/pods/359cbdf3-e6b5-44bf-9dc3-ec1dea8bf6e1/volumes" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.041951 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-config-data\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.042017 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7tsm4\" (UniqueName: \"kubernetes.io/projected/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-kube-api-access-7tsm4\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.042080 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.042201 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-scripts\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.042248 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-log-httpd\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.042421 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.042536 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-run-httpd\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.042878 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-log-httpd\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.043056 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-run-httpd\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.047426 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.047849 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.048913 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-config-data\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.060767 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-scripts\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.064420 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7tsm4\" (UniqueName: \"kubernetes.io/projected/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-kube-api-access-7tsm4\") pod \"ceilometer-0\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.176680 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:37:14 crc kubenswrapper[4930]: E0313 09:37:14.265596 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 09:37:14 crc kubenswrapper[4930]: E0313 09:37:14.288125 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 09:37:14 crc kubenswrapper[4930]: E0313 09:37:14.299076 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 09:37:14 crc kubenswrapper[4930]: E0313 09:37:14.299154 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-7f4b8495c9-mqbjk" podUID="7754b93c-5665-4f0d-8e26-d9910d1e90c4" containerName="heat-engine" Mar 13 09:37:14 crc kubenswrapper[4930]: I0313 09:37:14.840927 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.482364 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.584646 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-combined-ca-bundle\") pod \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.585055 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pxz6\" (UniqueName: \"kubernetes.io/projected/7754b93c-5665-4f0d-8e26-d9910d1e90c4-kube-api-access-7pxz6\") pod \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.585153 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-config-data\") pod \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.585557 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-config-data-custom\") pod \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\" (UID: \"7754b93c-5665-4f0d-8e26-d9910d1e90c4\") " Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.591591 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "7754b93c-5665-4f0d-8e26-d9910d1e90c4" (UID: "7754b93c-5665-4f0d-8e26-d9910d1e90c4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.591662 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7754b93c-5665-4f0d-8e26-d9910d1e90c4-kube-api-access-7pxz6" (OuterVolumeSpecName: "kube-api-access-7pxz6") pod "7754b93c-5665-4f0d-8e26-d9910d1e90c4" (UID: "7754b93c-5665-4f0d-8e26-d9910d1e90c4"). InnerVolumeSpecName "kube-api-access-7pxz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.639686 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7754b93c-5665-4f0d-8e26-d9910d1e90c4" (UID: "7754b93c-5665-4f0d-8e26-d9910d1e90c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.666620 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-config-data" (OuterVolumeSpecName: "config-data") pod "7754b93c-5665-4f0d-8e26-d9910d1e90c4" (UID: "7754b93c-5665-4f0d-8e26-d9910d1e90c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.688558 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pxz6\" (UniqueName: \"kubernetes.io/projected/7754b93c-5665-4f0d-8e26-d9910d1e90c4-kube-api-access-7pxz6\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.688591 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.688602 4930 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.688612 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7754b93c-5665-4f0d-8e26-d9910d1e90c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.806504 4930 generic.go:334] "Generic (PLEG): container finished" podID="7754b93c-5665-4f0d-8e26-d9910d1e90c4" containerID="db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e" exitCode=0 Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.806619 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-7f4b8495c9-mqbjk" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.806668 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7f4b8495c9-mqbjk" event={"ID":"7754b93c-5665-4f0d-8e26-d9910d1e90c4","Type":"ContainerDied","Data":"db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e"} Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.806716 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-7f4b8495c9-mqbjk" event={"ID":"7754b93c-5665-4f0d-8e26-d9910d1e90c4","Type":"ContainerDied","Data":"cbf3e3fbe0c39b5610758ce1592b65352077ac1dedf56f0fe4939c1a38009b9e"} Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.806736 4930 scope.go:117] "RemoveContainer" containerID="db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.808696 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce","Type":"ContainerStarted","Data":"b3ddc06004b4327d3fcf402aeac6dd92052ea51de69c75aad76b5ee7d2492982"} Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.808727 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce","Type":"ContainerStarted","Data":"a087db8074046bd0037a78648e96b1f92cf695ab2e94e283fbcef28812bce23f"} Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.861094 4930 scope.go:117] "RemoveContainer" containerID="db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e" Mar 13 09:37:15 crc kubenswrapper[4930]: E0313 09:37:15.861531 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e\": container with ID starting with db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e not found: ID does not exist" containerID="db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.861562 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e"} err="failed to get container status \"db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e\": rpc error: code = NotFound desc = could not find container \"db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e\": container with ID starting with db1f55fac0b88defee771f81887076b6e4855150208af73034c14586d001e14e not found: ID does not exist" Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.882483 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-7f4b8495c9-mqbjk"] Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.925533 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-7f4b8495c9-mqbjk"] Mar 13 09:37:15 crc kubenswrapper[4930]: I0313 09:37:15.983495 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7754b93c-5665-4f0d-8e26-d9910d1e90c4" path="/var/lib/kubelet/pods/7754b93c-5665-4f0d-8e26-d9910d1e90c4/volumes" Mar 13 09:37:16 crc kubenswrapper[4930]: I0313 09:37:16.825781 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce","Type":"ContainerStarted","Data":"9ea640efdbad2e911d61635f38085ac001ab59bac2821ecd82c95b6b76240f05"} Mar 13 09:37:17 crc kubenswrapper[4930]: I0313 09:37:17.842204 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce","Type":"ContainerStarted","Data":"d790e51bbff33aca0abec69815057106adf845af0526f40d6fa680d56233f359"} Mar 13 09:37:19 crc kubenswrapper[4930]: I0313 09:37:19.702343 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:19 crc kubenswrapper[4930]: I0313 09:37:19.864209 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce","Type":"ContainerStarted","Data":"1afb38383d504083635811002029a67363e961c9e2e7c7cc0d54df10d6b1c0a2"} Mar 13 09:37:19 crc kubenswrapper[4930]: I0313 09:37:19.864767 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 09:37:19 crc kubenswrapper[4930]: I0313 09:37:19.897792 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.587731498 podStartE2EDuration="6.89776989s" podCreationTimestamp="2026-03-13 09:37:13 +0000 UTC" firstStartedPulling="2026-03-13 09:37:14.872181037 +0000 UTC m=+1475.622095714" lastFinishedPulling="2026-03-13 09:37:19.182219429 +0000 UTC m=+1479.932134106" observedRunningTime="2026-03-13 09:37:19.885634752 +0000 UTC m=+1480.635549429" watchObservedRunningTime="2026-03-13 09:37:19.89776989 +0000 UTC m=+1480.647684557" Mar 13 09:37:20 crc kubenswrapper[4930]: I0313 09:37:20.900348 4930 generic.go:334] "Generic (PLEG): container finished" podID="f2a8264a-7145-4f0b-9f7e-4e9552dc84c0" containerID="93396c58766bfe9d2b827cfca22cf86170091e1e938e713952dfa6b7ed7f9479" exitCode=0 Mar 13 09:37:20 crc kubenswrapper[4930]: I0313 09:37:20.900930 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="ceilometer-central-agent" containerID="cri-o://b3ddc06004b4327d3fcf402aeac6dd92052ea51de69c75aad76b5ee7d2492982" gracePeriod=30 Mar 13 09:37:20 crc kubenswrapper[4930]: I0313 09:37:20.901008 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gb9wl" event={"ID":"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0","Type":"ContainerDied","Data":"93396c58766bfe9d2b827cfca22cf86170091e1e938e713952dfa6b7ed7f9479"} Mar 13 09:37:20 crc kubenswrapper[4930]: I0313 09:37:20.901636 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="proxy-httpd" containerID="cri-o://1afb38383d504083635811002029a67363e961c9e2e7c7cc0d54df10d6b1c0a2" gracePeriod=30 Mar 13 09:37:20 crc kubenswrapper[4930]: I0313 09:37:20.901702 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="sg-core" containerID="cri-o://d790e51bbff33aca0abec69815057106adf845af0526f40d6fa680d56233f359" gracePeriod=30 Mar 13 09:37:20 crc kubenswrapper[4930]: I0313 09:37:20.901763 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="ceilometer-notification-agent" containerID="cri-o://9ea640efdbad2e911d61635f38085ac001ab59bac2821ecd82c95b6b76240f05" gracePeriod=30 Mar 13 09:37:21 crc kubenswrapper[4930]: I0313 09:37:21.913990 4930 generic.go:334] "Generic (PLEG): container finished" podID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerID="1afb38383d504083635811002029a67363e961c9e2e7c7cc0d54df10d6b1c0a2" exitCode=0 Mar 13 09:37:21 crc kubenswrapper[4930]: I0313 09:37:21.914030 4930 generic.go:334] "Generic (PLEG): container finished" podID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerID="d790e51bbff33aca0abec69815057106adf845af0526f40d6fa680d56233f359" exitCode=2 Mar 13 09:37:21 crc kubenswrapper[4930]: I0313 09:37:21.914038 4930 generic.go:334] "Generic (PLEG): container finished" podID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerID="9ea640efdbad2e911d61635f38085ac001ab59bac2821ecd82c95b6b76240f05" exitCode=0 Mar 13 09:37:21 crc kubenswrapper[4930]: I0313 09:37:21.914243 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce","Type":"ContainerDied","Data":"1afb38383d504083635811002029a67363e961c9e2e7c7cc0d54df10d6b1c0a2"} Mar 13 09:37:21 crc kubenswrapper[4930]: I0313 09:37:21.914273 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce","Type":"ContainerDied","Data":"d790e51bbff33aca0abec69815057106adf845af0526f40d6fa680d56233f359"} Mar 13 09:37:21 crc kubenswrapper[4930]: I0313 09:37:21.914283 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce","Type":"ContainerDied","Data":"9ea640efdbad2e911d61635f38085ac001ab59bac2821ecd82c95b6b76240f05"} Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.389111 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.464110 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-config-data\") pod \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.464262 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-combined-ca-bundle\") pod \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.464372 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-scripts\") pod \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.464414 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzp98\" (UniqueName: \"kubernetes.io/projected/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-kube-api-access-xzp98\") pod \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\" (UID: \"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0\") " Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.476127 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-scripts" (OuterVolumeSpecName: "scripts") pod "f2a8264a-7145-4f0b-9f7e-4e9552dc84c0" (UID: "f2a8264a-7145-4f0b-9f7e-4e9552dc84c0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.498725 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-kube-api-access-xzp98" (OuterVolumeSpecName: "kube-api-access-xzp98") pod "f2a8264a-7145-4f0b-9f7e-4e9552dc84c0" (UID: "f2a8264a-7145-4f0b-9f7e-4e9552dc84c0"). InnerVolumeSpecName "kube-api-access-xzp98". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.506247 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f2a8264a-7145-4f0b-9f7e-4e9552dc84c0" (UID: "f2a8264a-7145-4f0b-9f7e-4e9552dc84c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.527866 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-config-data" (OuterVolumeSpecName: "config-data") pod "f2a8264a-7145-4f0b-9f7e-4e9552dc84c0" (UID: "f2a8264a-7145-4f0b-9f7e-4e9552dc84c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.567951 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.567989 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzp98\" (UniqueName: \"kubernetes.io/projected/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-kube-api-access-xzp98\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.568004 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.568014 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.926585 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-gb9wl" event={"ID":"f2a8264a-7145-4f0b-9f7e-4e9552dc84c0","Type":"ContainerDied","Data":"158e96e9538bce86be4b069370b744406c6584a19475b586d34622caa5f0aa64"} Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.926688 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="158e96e9538bce86be4b069370b744406c6584a19475b586d34622caa5f0aa64" Mar 13 09:37:22 crc kubenswrapper[4930]: I0313 09:37:22.926626 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-gb9wl" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.145939 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 09:37:23 crc kubenswrapper[4930]: E0313 09:37:23.146854 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f2a8264a-7145-4f0b-9f7e-4e9552dc84c0" containerName="nova-cell0-conductor-db-sync" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.146941 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f2a8264a-7145-4f0b-9f7e-4e9552dc84c0" containerName="nova-cell0-conductor-db-sync" Mar 13 09:37:23 crc kubenswrapper[4930]: E0313 09:37:23.147000 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7754b93c-5665-4f0d-8e26-d9910d1e90c4" containerName="heat-engine" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.156865 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="7754b93c-5665-4f0d-8e26-d9910d1e90c4" containerName="heat-engine" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.157597 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f2a8264a-7145-4f0b-9f7e-4e9552dc84c0" containerName="nova-cell0-conductor-db-sync" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.157700 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="7754b93c-5665-4f0d-8e26-d9910d1e90c4" containerName="heat-engine" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.159683 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.159894 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.162051 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-pdj49" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.163326 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.288484 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8\") " pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.289047 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8\") " pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.289557 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xh64z\" (UniqueName: \"kubernetes.io/projected/8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8-kube-api-access-xh64z\") pod \"nova-cell0-conductor-0\" (UID: \"8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8\") " pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.392032 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8\") " pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.392331 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8\") " pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.392583 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xh64z\" (UniqueName: \"kubernetes.io/projected/8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8-kube-api-access-xh64z\") pod \"nova-cell0-conductor-0\" (UID: \"8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8\") " pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.397710 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8\") " pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.398463 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8\") " pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.416537 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xh64z\" (UniqueName: \"kubernetes.io/projected/8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8-kube-api-access-xh64z\") pod \"nova-cell0-conductor-0\" (UID: \"8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8\") " pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:23 crc kubenswrapper[4930]: I0313 09:37:23.485721 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:24 crc kubenswrapper[4930]: I0313 09:37:24.079718 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 13 09:37:24 crc kubenswrapper[4930]: I0313 09:37:24.967923 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8","Type":"ContainerStarted","Data":"5dbacc1057d2f372063cf04b4d76172deddfc37adea603e94b2024f9b1b08324"} Mar 13 09:37:24 crc kubenswrapper[4930]: I0313 09:37:24.968237 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"8fbd5b6b-8890-4f7b-bad5-a7c6de7f60a8","Type":"ContainerStarted","Data":"4fcdd2bf2f6c551d1264963124ecd8ee5277ef4451b87a1674f211837fdba323"} Mar 13 09:37:24 crc kubenswrapper[4930]: I0313 09:37:24.968527 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:24 crc kubenswrapper[4930]: I0313 09:37:24.972204 4930 generic.go:334] "Generic (PLEG): container finished" podID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerID="b3ddc06004b4327d3fcf402aeac6dd92052ea51de69c75aad76b5ee7d2492982" exitCode=0 Mar 13 09:37:24 crc kubenswrapper[4930]: I0313 09:37:24.972245 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce","Type":"ContainerDied","Data":"b3ddc06004b4327d3fcf402aeac6dd92052ea51de69c75aad76b5ee7d2492982"} Mar 13 09:37:24 crc kubenswrapper[4930]: I0313 09:37:24.991049 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=1.991028063 podStartE2EDuration="1.991028063s" podCreationTimestamp="2026-03-13 09:37:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:37:24.98785852 +0000 UTC m=+1485.737773197" watchObservedRunningTime="2026-03-13 09:37:24.991028063 +0000 UTC m=+1485.740942740" Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.728100 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.856025 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-scripts\") pod \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.856186 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-config-data\") pod \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.856221 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-combined-ca-bundle\") pod \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.856342 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-run-httpd\") pod \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.856490 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-log-httpd\") pod \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.856514 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-sg-core-conf-yaml\") pod \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.856614 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7tsm4\" (UniqueName: \"kubernetes.io/projected/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-kube-api-access-7tsm4\") pod \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\" (UID: \"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce\") " Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.857299 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" (UID: "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.857596 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" (UID: "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.878711 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-kube-api-access-7tsm4" (OuterVolumeSpecName: "kube-api-access-7tsm4") pod "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" (UID: "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce"). InnerVolumeSpecName "kube-api-access-7tsm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.880645 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-scripts" (OuterVolumeSpecName: "scripts") pod "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" (UID: "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.946130 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" (UID: "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.959731 4930 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.959765 4930 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.959776 4930 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.959800 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7tsm4\" (UniqueName: \"kubernetes.io/projected/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-kube-api-access-7tsm4\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:25 crc kubenswrapper[4930]: I0313 09:37:25.959809 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.005095 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" (UID: "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.006304 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.062320 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.064945 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-config-data" (OuterVolumeSpecName: "config-data") pod "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" (UID: "dc77f031-61bc-4486-a0ea-db6f4e8ed2ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.109388 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"dc77f031-61bc-4486-a0ea-db6f4e8ed2ce","Type":"ContainerDied","Data":"a087db8074046bd0037a78648e96b1f92cf695ab2e94e283fbcef28812bce23f"} Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.109456 4930 scope.go:117] "RemoveContainer" containerID="1afb38383d504083635811002029a67363e961c9e2e7c7cc0d54df10d6b1c0a2" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.141034 4930 scope.go:117] "RemoveContainer" containerID="d790e51bbff33aca0abec69815057106adf845af0526f40d6fa680d56233f359" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.166506 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.170877 4930 scope.go:117] "RemoveContainer" containerID="9ea640efdbad2e911d61635f38085ac001ab59bac2821ecd82c95b6b76240f05" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.195125 4930 scope.go:117] "RemoveContainer" containerID="b3ddc06004b4327d3fcf402aeac6dd92052ea51de69c75aad76b5ee7d2492982" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.352653 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.372330 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.394477 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:26 crc kubenswrapper[4930]: E0313 09:37:26.394969 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="ceilometer-central-agent" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.394991 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="ceilometer-central-agent" Mar 13 09:37:26 crc kubenswrapper[4930]: E0313 09:37:26.394999 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="ceilometer-notification-agent" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.395005 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="ceilometer-notification-agent" Mar 13 09:37:26 crc kubenswrapper[4930]: E0313 09:37:26.395022 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="proxy-httpd" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.395028 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="proxy-httpd" Mar 13 09:37:26 crc kubenswrapper[4930]: E0313 09:37:26.395050 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="sg-core" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.395109 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="sg-core" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.395390 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="ceilometer-central-agent" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.395409 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="proxy-httpd" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.395422 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="ceilometer-notification-agent" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.395476 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" containerName="sg-core" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.398326 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.401225 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.402728 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.409281 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.473447 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-config-data\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.473608 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwh7t\" (UniqueName: \"kubernetes.io/projected/e40e80e4-7348-4e09-b661-c0c55c8c4a18-kube-api-access-bwh7t\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.473637 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e40e80e4-7348-4e09-b661-c0c55c8c4a18-run-httpd\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.473669 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.473808 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e40e80e4-7348-4e09-b661-c0c55c8c4a18-log-httpd\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.473858 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-scripts\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.473887 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.575681 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-config-data\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.575786 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwh7t\" (UniqueName: \"kubernetes.io/projected/e40e80e4-7348-4e09-b661-c0c55c8c4a18-kube-api-access-bwh7t\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.575814 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e40e80e4-7348-4e09-b661-c0c55c8c4a18-run-httpd\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.575840 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.575941 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e40e80e4-7348-4e09-b661-c0c55c8c4a18-log-httpd\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.575977 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-scripts\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.575997 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.576661 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e40e80e4-7348-4e09-b661-c0c55c8c4a18-log-httpd\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.576832 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e40e80e4-7348-4e09-b661-c0c55c8c4a18-run-httpd\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.580537 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.582153 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-scripts\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.582236 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-config-data\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.584222 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.599426 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwh7t\" (UniqueName: \"kubernetes.io/projected/e40e80e4-7348-4e09-b661-c0c55c8c4a18-kube-api-access-bwh7t\") pod \"ceilometer-0\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " pod="openstack/ceilometer-0" Mar 13 09:37:26 crc kubenswrapper[4930]: I0313 09:37:26.723745 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:37:27 crc kubenswrapper[4930]: I0313 09:37:27.230312 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.031082 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e40e80e4-7348-4e09-b661-c0c55c8c4a18","Type":"ContainerStarted","Data":"516962bc1d9e7b69a9366483f9ee70c6f5acd20b3b46c1c003409ebf2725cf7c"} Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.122722 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc77f031-61bc-4486-a0ea-db6f4e8ed2ce" path="/var/lib/kubelet/pods/dc77f031-61bc-4486-a0ea-db6f4e8ed2ce/volumes" Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.757810 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-95hs2"] Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.770727 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.807419 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-95hs2"] Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.843050 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a86863c-63e1-4921-bd79-e6543a560043-utilities\") pod \"certified-operators-95hs2\" (UID: \"0a86863c-63e1-4921-bd79-e6543a560043\") " pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.843458 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a86863c-63e1-4921-bd79-e6543a560043-catalog-content\") pod \"certified-operators-95hs2\" (UID: \"0a86863c-63e1-4921-bd79-e6543a560043\") " pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.843700 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjkkv\" (UniqueName: \"kubernetes.io/projected/0a86863c-63e1-4921-bd79-e6543a560043-kube-api-access-xjkkv\") pod \"certified-operators-95hs2\" (UID: \"0a86863c-63e1-4921-bd79-e6543a560043\") " pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.945112 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a86863c-63e1-4921-bd79-e6543a560043-catalog-content\") pod \"certified-operators-95hs2\" (UID: \"0a86863c-63e1-4921-bd79-e6543a560043\") " pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.945221 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjkkv\" (UniqueName: \"kubernetes.io/projected/0a86863c-63e1-4921-bd79-e6543a560043-kube-api-access-xjkkv\") pod \"certified-operators-95hs2\" (UID: \"0a86863c-63e1-4921-bd79-e6543a560043\") " pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.945276 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a86863c-63e1-4921-bd79-e6543a560043-utilities\") pod \"certified-operators-95hs2\" (UID: \"0a86863c-63e1-4921-bd79-e6543a560043\") " pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.945717 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a86863c-63e1-4921-bd79-e6543a560043-catalog-content\") pod \"certified-operators-95hs2\" (UID: \"0a86863c-63e1-4921-bd79-e6543a560043\") " pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.945758 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a86863c-63e1-4921-bd79-e6543a560043-utilities\") pod \"certified-operators-95hs2\" (UID: \"0a86863c-63e1-4921-bd79-e6543a560043\") " pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:28 crc kubenswrapper[4930]: I0313 09:37:28.973630 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjkkv\" (UniqueName: \"kubernetes.io/projected/0a86863c-63e1-4921-bd79-e6543a560043-kube-api-access-xjkkv\") pod \"certified-operators-95hs2\" (UID: \"0a86863c-63e1-4921-bd79-e6543a560043\") " pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:29 crc kubenswrapper[4930]: I0313 09:37:29.043584 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e40e80e4-7348-4e09-b661-c0c55c8c4a18","Type":"ContainerStarted","Data":"1531e381efb54d1fc2d4bc432da7030534b007725ca21cf8e228d57f89d37dc5"} Mar 13 09:37:29 crc kubenswrapper[4930]: I0313 09:37:29.260612 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:29 crc kubenswrapper[4930]: I0313 09:37:29.767802 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-95hs2"] Mar 13 09:37:30 crc kubenswrapper[4930]: I0313 09:37:30.030547 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:30 crc kubenswrapper[4930]: I0313 09:37:30.073239 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e40e80e4-7348-4e09-b661-c0c55c8c4a18","Type":"ContainerStarted","Data":"fa7671ba97dda70ce688b6467bf102a454c07a2d07eb49426cd807b252b1c567"} Mar 13 09:37:30 crc kubenswrapper[4930]: I0313 09:37:30.074313 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-95hs2" event={"ID":"0a86863c-63e1-4921-bd79-e6543a560043","Type":"ContainerStarted","Data":"b24a76b4c42bac10ef90354044d285777921d357d0cb1a27ccedd02f8471bdc3"} Mar 13 09:37:31 crc kubenswrapper[4930]: I0313 09:37:31.087026 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e40e80e4-7348-4e09-b661-c0c55c8c4a18","Type":"ContainerStarted","Data":"79f3124ee3849ab23fc945701f13ca3bee7f4016aa170a922bd6d7adc3d23179"} Mar 13 09:37:31 crc kubenswrapper[4930]: I0313 09:37:31.089808 4930 generic.go:334] "Generic (PLEG): container finished" podID="0a86863c-63e1-4921-bd79-e6543a560043" containerID="c42cbb1bbad0aee1ec5f69e1ac384be8a2a188a84624e825bf5f4748d1d47540" exitCode=0 Mar 13 09:37:31 crc kubenswrapper[4930]: I0313 09:37:31.089861 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-95hs2" event={"ID":"0a86863c-63e1-4921-bd79-e6543a560043","Type":"ContainerDied","Data":"c42cbb1bbad0aee1ec5f69e1ac384be8a2a188a84624e825bf5f4748d1d47540"} Mar 13 09:37:33 crc kubenswrapper[4930]: I0313 09:37:33.129906 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e40e80e4-7348-4e09-b661-c0c55c8c4a18","Type":"ContainerStarted","Data":"035f8969f11fbec2c070d2865cb2f2b53e02519f631b9e537036f3df8bbcc70c"} Mar 13 09:37:33 crc kubenswrapper[4930]: I0313 09:37:33.130618 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 09:37:33 crc kubenswrapper[4930]: I0313 09:37:33.130530 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="sg-core" containerID="cri-o://79f3124ee3849ab23fc945701f13ca3bee7f4016aa170a922bd6d7adc3d23179" gracePeriod=30 Mar 13 09:37:33 crc kubenswrapper[4930]: I0313 09:37:33.130008 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="ceilometer-central-agent" containerID="cri-o://1531e381efb54d1fc2d4bc432da7030534b007725ca21cf8e228d57f89d37dc5" gracePeriod=30 Mar 13 09:37:33 crc kubenswrapper[4930]: I0313 09:37:33.130545 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="proxy-httpd" containerID="cri-o://035f8969f11fbec2c070d2865cb2f2b53e02519f631b9e537036f3df8bbcc70c" gracePeriod=30 Mar 13 09:37:33 crc kubenswrapper[4930]: I0313 09:37:33.130560 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="ceilometer-notification-agent" containerID="cri-o://fa7671ba97dda70ce688b6467bf102a454c07a2d07eb49426cd807b252b1c567" gracePeriod=30 Mar 13 09:37:33 crc kubenswrapper[4930]: I0313 09:37:33.150405 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-95hs2" event={"ID":"0a86863c-63e1-4921-bd79-e6543a560043","Type":"ContainerStarted","Data":"e8a12b74600495ac5afdbfa941792d415a79d4f129c347f079f29fe6d6d02978"} Mar 13 09:37:33 crc kubenswrapper[4930]: I0313 09:37:33.178839 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.578101241 podStartE2EDuration="7.178818761s" podCreationTimestamp="2026-03-13 09:37:26 +0000 UTC" firstStartedPulling="2026-03-13 09:37:27.234576232 +0000 UTC m=+1487.984490909" lastFinishedPulling="2026-03-13 09:37:31.835293752 +0000 UTC m=+1492.585208429" observedRunningTime="2026-03-13 09:37:33.161157629 +0000 UTC m=+1493.911072306" watchObservedRunningTime="2026-03-13 09:37:33.178818761 +0000 UTC m=+1493.928733438" Mar 13 09:37:33 crc kubenswrapper[4930]: I0313 09:37:33.525258 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 13 09:37:33 crc kubenswrapper[4930]: E0313 09:37:33.590257 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode40e80e4_7348_4e09_b661_c0c55c8c4a18.slice/crio-79f3124ee3849ab23fc945701f13ca3bee7f4016aa170a922bd6d7adc3d23179.scope\": RecentStats: unable to find data in memory cache]" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.141384 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-hljq9"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.144245 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.149709 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.149986 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.164797 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-hljq9"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.169196 4930 generic.go:334] "Generic (PLEG): container finished" podID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerID="035f8969f11fbec2c070d2865cb2f2b53e02519f631b9e537036f3df8bbcc70c" exitCode=0 Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.169232 4930 generic.go:334] "Generic (PLEG): container finished" podID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerID="79f3124ee3849ab23fc945701f13ca3bee7f4016aa170a922bd6d7adc3d23179" exitCode=2 Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.169258 4930 generic.go:334] "Generic (PLEG): container finished" podID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerID="fa7671ba97dda70ce688b6467bf102a454c07a2d07eb49426cd807b252b1c567" exitCode=0 Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.170374 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e40e80e4-7348-4e09-b661-c0c55c8c4a18","Type":"ContainerDied","Data":"035f8969f11fbec2c070d2865cb2f2b53e02519f631b9e537036f3df8bbcc70c"} Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.170415 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e40e80e4-7348-4e09-b661-c0c55c8c4a18","Type":"ContainerDied","Data":"79f3124ee3849ab23fc945701f13ca3bee7f4016aa170a922bd6d7adc3d23179"} Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.170451 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e40e80e4-7348-4e09-b661-c0c55c8c4a18","Type":"ContainerDied","Data":"fa7671ba97dda70ce688b6467bf102a454c07a2d07eb49426cd807b252b1c567"} Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.222921 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2zs2\" (UniqueName: \"kubernetes.io/projected/67ceb05f-14a6-4adb-af90-5d52549e4f23-kube-api-access-j2zs2\") pod \"nova-cell0-cell-mapping-hljq9\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.223093 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hljq9\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.223125 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-config-data\") pod \"nova-cell0-cell-mapping-hljq9\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.223179 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-scripts\") pod \"nova-cell0-cell-mapping-hljq9\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.325174 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2zs2\" (UniqueName: \"kubernetes.io/projected/67ceb05f-14a6-4adb-af90-5d52549e4f23-kube-api-access-j2zs2\") pod \"nova-cell0-cell-mapping-hljq9\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.325511 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hljq9\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.325573 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-config-data\") pod \"nova-cell0-cell-mapping-hljq9\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.325632 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-scripts\") pod \"nova-cell0-cell-mapping-hljq9\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.339163 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-config-data\") pod \"nova-cell0-cell-mapping-hljq9\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.362909 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-scripts\") pod \"nova-cell0-cell-mapping-hljq9\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.363483 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-hljq9\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.385330 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2zs2\" (UniqueName: \"kubernetes.io/projected/67ceb05f-14a6-4adb-af90-5d52549e4f23-kube-api-access-j2zs2\") pod \"nova-cell0-cell-mapping-hljq9\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.429535 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.431477 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.450950 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.464964 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.481001 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-r7nvk"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.491001 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-r7nvk" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.535180 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3fee40-f7de-42df-b078-abde4117d47a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.535461 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7bz6\" (UniqueName: \"kubernetes.io/projected/05b2d756-3fdc-4914-b511-350ee81fabb4-kube-api-access-n7bz6\") pod \"aodh-db-create-r7nvk\" (UID: \"05b2d756-3fdc-4914-b511-350ee81fabb4\") " pod="openstack/aodh-db-create-r7nvk" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.535552 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3fee40-f7de-42df-b078-abde4117d47a-config-data\") pod \"nova-api-0\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.535627 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05b2d756-3fdc-4914-b511-350ee81fabb4-operator-scripts\") pod \"aodh-db-create-r7nvk\" (UID: \"05b2d756-3fdc-4914-b511-350ee81fabb4\") " pod="openstack/aodh-db-create-r7nvk" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.535707 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a3fee40-f7de-42df-b078-abde4117d47a-logs\") pod \"nova-api-0\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.535773 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xljc6\" (UniqueName: \"kubernetes.io/projected/1a3fee40-f7de-42df-b078-abde4117d47a-kube-api-access-xljc6\") pod \"nova-api-0\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.578133 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.641774 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3fee40-f7de-42df-b078-abde4117d47a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.641881 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7bz6\" (UniqueName: \"kubernetes.io/projected/05b2d756-3fdc-4914-b511-350ee81fabb4-kube-api-access-n7bz6\") pod \"aodh-db-create-r7nvk\" (UID: \"05b2d756-3fdc-4914-b511-350ee81fabb4\") " pod="openstack/aodh-db-create-r7nvk" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.641909 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3fee40-f7de-42df-b078-abde4117d47a-config-data\") pod \"nova-api-0\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.641931 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05b2d756-3fdc-4914-b511-350ee81fabb4-operator-scripts\") pod \"aodh-db-create-r7nvk\" (UID: \"05b2d756-3fdc-4914-b511-350ee81fabb4\") " pod="openstack/aodh-db-create-r7nvk" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.641977 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a3fee40-f7de-42df-b078-abde4117d47a-logs\") pod \"nova-api-0\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.641997 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xljc6\" (UniqueName: \"kubernetes.io/projected/1a3fee40-f7de-42df-b078-abde4117d47a-kube-api-access-xljc6\") pod \"nova-api-0\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.643225 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05b2d756-3fdc-4914-b511-350ee81fabb4-operator-scripts\") pod \"aodh-db-create-r7nvk\" (UID: \"05b2d756-3fdc-4914-b511-350ee81fabb4\") " pod="openstack/aodh-db-create-r7nvk" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.643578 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a3fee40-f7de-42df-b078-abde4117d47a-logs\") pod \"nova-api-0\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.665677 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-r7nvk"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.695804 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3fee40-f7de-42df-b078-abde4117d47a-config-data\") pod \"nova-api-0\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.720568 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xljc6\" (UniqueName: \"kubernetes.io/projected/1a3fee40-f7de-42df-b078-abde4117d47a-kube-api-access-xljc6\") pod \"nova-api-0\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.739581 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7bz6\" (UniqueName: \"kubernetes.io/projected/05b2d756-3fdc-4914-b511-350ee81fabb4-kube-api-access-n7bz6\") pod \"aodh-db-create-r7nvk\" (UID: \"05b2d756-3fdc-4914-b511-350ee81fabb4\") " pod="openstack/aodh-db-create-r7nvk" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.741171 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3fee40-f7de-42df-b078-abde4117d47a-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.783569 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.785627 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.789279 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.804954 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.807443 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.807488 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.810385 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.822709 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-r7nvk" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.832876 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.858921 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.872684 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.872759 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.873517 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " pod="openstack/nova-metadata-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.873643 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-logs\") pod \"nova-metadata-0\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " pod="openstack/nova-metadata-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.873728 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-config-data\") pod \"nova-metadata-0\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " pod="openstack/nova-metadata-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.873856 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7rf9\" (UniqueName: \"kubernetes.io/projected/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-kube-api-access-w7rf9\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.874027 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92p2s\" (UniqueName: \"kubernetes.io/projected/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-kube-api-access-92p2s\") pod \"nova-metadata-0\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " pod="openstack/nova-metadata-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.901368 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-3a0f-account-create-update-82l86"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.903724 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-3a0f-account-create-update-82l86" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.908714 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.950229 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.970193 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.971396 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.975724 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.979172 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7rf9\" (UniqueName: \"kubernetes.io/projected/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-kube-api-access-w7rf9\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.979479 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mstdv\" (UniqueName: \"kubernetes.io/projected/e5c00fcf-72f0-45e6-a1b3-235677ec4d14-kube-api-access-mstdv\") pod \"aodh-3a0f-account-create-update-82l86\" (UID: \"e5c00fcf-72f0-45e6-a1b3-235677ec4d14\") " pod="openstack/aodh-3a0f-account-create-update-82l86" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.980406 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92p2s\" (UniqueName: \"kubernetes.io/projected/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-kube-api-access-92p2s\") pod \"nova-metadata-0\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " pod="openstack/nova-metadata-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.980946 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.981025 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.981524 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " pod="openstack/nova-metadata-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.981599 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5c00fcf-72f0-45e6-a1b3-235677ec4d14-operator-scripts\") pod \"aodh-3a0f-account-create-update-82l86\" (UID: \"e5c00fcf-72f0-45e6-a1b3-235677ec4d14\") " pod="openstack/aodh-3a0f-account-create-update-82l86" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.981922 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-logs\") pod \"nova-metadata-0\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " pod="openstack/nova-metadata-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.982006 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-config-data\") pod \"nova-metadata-0\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " pod="openstack/nova-metadata-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.985692 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-logs\") pod \"nova-metadata-0\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " pod="openstack/nova-metadata-0" Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.986677 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-3a0f-account-create-update-82l86"] Mar 13 09:37:34 crc kubenswrapper[4930]: I0313 09:37:34.988496 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.003158 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-config-data\") pod \"nova-metadata-0\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " pod="openstack/nova-metadata-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.003372 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-t47jv"] Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.005425 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.006362 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.006723 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " pod="openstack/nova-metadata-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.024253 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92p2s\" (UniqueName: \"kubernetes.io/projected/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-kube-api-access-92p2s\") pod \"nova-metadata-0\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " pod="openstack/nova-metadata-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.032560 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-t47jv"] Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.035715 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7rf9\" (UniqueName: \"kubernetes.io/projected/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-kube-api-access-w7rf9\") pod \"nova-cell1-novncproxy-0\" (UID: \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.084717 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5c00fcf-72f0-45e6-a1b3-235677ec4d14-operator-scripts\") pod \"aodh-3a0f-account-create-update-82l86\" (UID: \"e5c00fcf-72f0-45e6-a1b3-235677ec4d14\") " pod="openstack/aodh-3a0f-account-create-update-82l86" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.085069 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.085138 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b7ebf6b-9957-4833-be7b-2123a7a456ee-config-data\") pod \"nova-scheduler-0\" (UID: \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.085205 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.085231 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvctp\" (UniqueName: \"kubernetes.io/projected/4b7ebf6b-9957-4833-be7b-2123a7a456ee-kube-api-access-vvctp\") pod \"nova-scheduler-0\" (UID: \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.085259 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bbwr\" (UniqueName: \"kubernetes.io/projected/ea17f053-2c9a-42ba-b729-106a9f9b90bb-kube-api-access-4bbwr\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.085294 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.085367 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mstdv\" (UniqueName: \"kubernetes.io/projected/e5c00fcf-72f0-45e6-a1b3-235677ec4d14-kube-api-access-mstdv\") pod \"aodh-3a0f-account-create-update-82l86\" (UID: \"e5c00fcf-72f0-45e6-a1b3-235677ec4d14\") " pod="openstack/aodh-3a0f-account-create-update-82l86" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.085448 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.085478 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7ebf6b-9957-4833-be7b-2123a7a456ee-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.085507 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-config\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.086793 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5c00fcf-72f0-45e6-a1b3-235677ec4d14-operator-scripts\") pod \"aodh-3a0f-account-create-update-82l86\" (UID: \"e5c00fcf-72f0-45e6-a1b3-235677ec4d14\") " pod="openstack/aodh-3a0f-account-create-update-82l86" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.111256 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mstdv\" (UniqueName: \"kubernetes.io/projected/e5c00fcf-72f0-45e6-a1b3-235677ec4d14-kube-api-access-mstdv\") pod \"aodh-3a0f-account-create-update-82l86\" (UID: \"e5c00fcf-72f0-45e6-a1b3-235677ec4d14\") " pod="openstack/aodh-3a0f-account-create-update-82l86" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.133065 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.160061 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.188116 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.188185 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b7ebf6b-9957-4833-be7b-2123a7a456ee-config-data\") pod \"nova-scheduler-0\" (UID: \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.188228 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.188246 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvctp\" (UniqueName: \"kubernetes.io/projected/4b7ebf6b-9957-4833-be7b-2123a7a456ee-kube-api-access-vvctp\") pod \"nova-scheduler-0\" (UID: \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.188269 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bbwr\" (UniqueName: \"kubernetes.io/projected/ea17f053-2c9a-42ba-b729-106a9f9b90bb-kube-api-access-4bbwr\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.188295 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.188368 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7ebf6b-9957-4833-be7b-2123a7a456ee-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.188385 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.188406 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-config\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.189713 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-dns-swift-storage-0\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.189947 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-ovsdbserver-nb\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.190164 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-ovsdbserver-sb\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.190467 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-dns-svc\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.192075 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-config\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.195141 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b7ebf6b-9957-4833-be7b-2123a7a456ee-config-data\") pod \"nova-scheduler-0\" (UID: \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.195912 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7ebf6b-9957-4833-be7b-2123a7a456ee-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.206595 4930 generic.go:334] "Generic (PLEG): container finished" podID="0a86863c-63e1-4921-bd79-e6543a560043" containerID="e8a12b74600495ac5afdbfa941792d415a79d4f129c347f079f29fe6d6d02978" exitCode=0 Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.206649 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-95hs2" event={"ID":"0a86863c-63e1-4921-bd79-e6543a560043","Type":"ContainerDied","Data":"e8a12b74600495ac5afdbfa941792d415a79d4f129c347f079f29fe6d6d02978"} Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.211304 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bbwr\" (UniqueName: \"kubernetes.io/projected/ea17f053-2c9a-42ba-b729-106a9f9b90bb-kube-api-access-4bbwr\") pod \"dnsmasq-dns-568d7fd7cf-t47jv\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.216345 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvctp\" (UniqueName: \"kubernetes.io/projected/4b7ebf6b-9957-4833-be7b-2123a7a456ee-kube-api-access-vvctp\") pod \"nova-scheduler-0\" (UID: \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.244506 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-3a0f-account-create-update-82l86" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.326668 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.356594 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.453855 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-hljq9"] Mar 13 09:37:35 crc kubenswrapper[4930]: W0313 09:37:35.516479 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67ceb05f_14a6_4adb_af90_5d52549e4f23.slice/crio-0af727263de5b4a55896b7077a62e63e28522d786577b7eb02b7df3c951b3467 WatchSource:0}: Error finding container 0af727263de5b4a55896b7077a62e63e28522d786577b7eb02b7df3c951b3467: Status 404 returned error can't find the container with id 0af727263de5b4a55896b7077a62e63e28522d786577b7eb02b7df3c951b3467 Mar 13 09:37:35 crc kubenswrapper[4930]: I0313 09:37:35.680102 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:37:35 crc kubenswrapper[4930]: W0313 09:37:35.700080 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a3fee40_f7de_42df_b078_abde4117d47a.slice/crio-030358db98d8e20d07411e9b9d1a3d94f161410d3a163098669a11735612c7bb WatchSource:0}: Error finding container 030358db98d8e20d07411e9b9d1a3d94f161410d3a163098669a11735612c7bb: Status 404 returned error can't find the container with id 030358db98d8e20d07411e9b9d1a3d94f161410d3a163098669a11735612c7bb Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.169593 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-r7nvk"] Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.209825 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 09:37:36 crc kubenswrapper[4930]: W0313 09:37:36.214531 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfb80ecb_f616_4aa9_9eb4_3e17e2f932cb.slice/crio-5c81fe72845fc14f5c6931fc16a77d95a67bfe25bd3d904c1ee6aac840aa3c93 WatchSource:0}: Error finding container 5c81fe72845fc14f5c6931fc16a77d95a67bfe25bd3d904c1ee6aac840aa3c93: Status 404 returned error can't find the container with id 5c81fe72845fc14f5c6931fc16a77d95a67bfe25bd3d904c1ee6aac840aa3c93 Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.254523 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-95hs2" event={"ID":"0a86863c-63e1-4921-bd79-e6543a560043","Type":"ContainerStarted","Data":"6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be"} Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.277026 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1a3fee40-f7de-42df-b078-abde4117d47a","Type":"ContainerStarted","Data":"030358db98d8e20d07411e9b9d1a3d94f161410d3a163098669a11735612c7bb"} Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.282971 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hljq9" event={"ID":"67ceb05f-14a6-4adb-af90-5d52549e4f23","Type":"ContainerStarted","Data":"3b32b5ccd512cccbb3e04056bfad8de1f17b714917b17aac986da63abd41b2f0"} Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.290544 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hljq9" event={"ID":"67ceb05f-14a6-4adb-af90-5d52549e4f23","Type":"ContainerStarted","Data":"0af727263de5b4a55896b7077a62e63e28522d786577b7eb02b7df3c951b3467"} Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.379794 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-95hs2" podStartSLOduration=3.67671494 podStartE2EDuration="8.379772237s" podCreationTimestamp="2026-03-13 09:37:28 +0000 UTC" firstStartedPulling="2026-03-13 09:37:31.091898903 +0000 UTC m=+1491.841813580" lastFinishedPulling="2026-03-13 09:37:35.7949562 +0000 UTC m=+1496.544870877" observedRunningTime="2026-03-13 09:37:36.297310254 +0000 UTC m=+1497.047224951" watchObservedRunningTime="2026-03-13 09:37:36.379772237 +0000 UTC m=+1497.129686924" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.397072 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-hljq9" podStartSLOduration=2.397051616 podStartE2EDuration="2.397051616s" podCreationTimestamp="2026-03-13 09:37:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:37:36.32040721 +0000 UTC m=+1497.070321897" watchObservedRunningTime="2026-03-13 09:37:36.397051616 +0000 UTC m=+1497.146966293" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.559051 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.602287 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8pbts"] Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.603815 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.615059 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.615401 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.631420 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-3a0f-account-create-update-82l86"] Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.660756 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8pbts"] Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.698219 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-config-data\") pod \"nova-cell1-conductor-db-sync-8pbts\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.699853 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxn67\" (UniqueName: \"kubernetes.io/projected/7ccda94d-acb1-4c49-91e1-db24243cf6a9-kube-api-access-qxn67\") pod \"nova-cell1-conductor-db-sync-8pbts\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.699979 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-scripts\") pod \"nova-cell1-conductor-db-sync-8pbts\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.700659 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8pbts\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.802835 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8pbts\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.803186 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-config-data\") pod \"nova-cell1-conductor-db-sync-8pbts\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.803212 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxn67\" (UniqueName: \"kubernetes.io/projected/7ccda94d-acb1-4c49-91e1-db24243cf6a9-kube-api-access-qxn67\") pod \"nova-cell1-conductor-db-sync-8pbts\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.803631 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-scripts\") pod \"nova-cell1-conductor-db-sync-8pbts\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.811390 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-8pbts\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.820974 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-config-data\") pod \"nova-cell1-conductor-db-sync-8pbts\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.824729 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-scripts\") pod \"nova-cell1-conductor-db-sync-8pbts\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.824741 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxn67\" (UniqueName: \"kubernetes.io/projected/7ccda94d-acb1-4c49-91e1-db24243cf6a9-kube-api-access-qxn67\") pod \"nova-cell1-conductor-db-sync-8pbts\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.957546 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:37:36 crc kubenswrapper[4930]: I0313 09:37:36.987256 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-t47jv"] Mar 13 09:37:37 crc kubenswrapper[4930]: I0313 09:37:37.043345 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:37 crc kubenswrapper[4930]: I0313 09:37:37.339850 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" event={"ID":"ea17f053-2c9a-42ba-b729-106a9f9b90bb","Type":"ContainerStarted","Data":"85c6a2bdea7e7230f6517d2c8c446febf96489e1f11f49ce94f527301039a663"} Mar 13 09:37:37 crc kubenswrapper[4930]: I0313 09:37:37.354727 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b1d06be6-d50e-4bc3-8579-7a97b48f32b2","Type":"ContainerStarted","Data":"f208b782d2543d8e08baa20ea5215802363034935a2e9a8b6e5645efa1c6e0b2"} Mar 13 09:37:37 crc kubenswrapper[4930]: I0313 09:37:37.356852 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-3a0f-account-create-update-82l86" event={"ID":"e5c00fcf-72f0-45e6-a1b3-235677ec4d14","Type":"ContainerStarted","Data":"e55d97b040c4f9dd7afc52670411587db48d28ec123410d9c0cd8f8198000fc7"} Mar 13 09:37:37 crc kubenswrapper[4930]: I0313 09:37:37.356874 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-3a0f-account-create-update-82l86" event={"ID":"e5c00fcf-72f0-45e6-a1b3-235677ec4d14","Type":"ContainerStarted","Data":"b0a53bef955941e2fce6f29f916d75e3e7df3db60a28b461597f5b2379af698c"} Mar 13 09:37:37 crc kubenswrapper[4930]: I0313 09:37:37.358659 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4b7ebf6b-9957-4833-be7b-2123a7a456ee","Type":"ContainerStarted","Data":"c82dd72b332f72fdbb6883605d8f6273bfce4c53b9907e57f19ef3bc870e27c0"} Mar 13 09:37:37 crc kubenswrapper[4930]: I0313 09:37:37.360747 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb","Type":"ContainerStarted","Data":"5c81fe72845fc14f5c6931fc16a77d95a67bfe25bd3d904c1ee6aac840aa3c93"} Mar 13 09:37:37 crc kubenswrapper[4930]: I0313 09:37:37.374569 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-r7nvk" event={"ID":"05b2d756-3fdc-4914-b511-350ee81fabb4","Type":"ContainerStarted","Data":"7c2d12ac50c8e12bd82ecd79c65662aa39860ec7c3c0512ad6f6716b34b3987d"} Mar 13 09:37:37 crc kubenswrapper[4930]: I0313 09:37:37.374632 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-r7nvk" event={"ID":"05b2d756-3fdc-4914-b511-350ee81fabb4","Type":"ContainerStarted","Data":"7d0cca45d010cf2e7fbc97b2f83bbcb5f792119dfafc3348e727c37faca08a96"} Mar 13 09:37:37 crc kubenswrapper[4930]: I0313 09:37:37.408767 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-3a0f-account-create-update-82l86" podStartSLOduration=3.408739791 podStartE2EDuration="3.408739791s" podCreationTimestamp="2026-03-13 09:37:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:37:37.378631287 +0000 UTC m=+1498.128545964" watchObservedRunningTime="2026-03-13 09:37:37.408739791 +0000 UTC m=+1498.158654468" Mar 13 09:37:37 crc kubenswrapper[4930]: I0313 09:37:37.424598 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-create-r7nvk" podStartSLOduration=3.424578004 podStartE2EDuration="3.424578004s" podCreationTimestamp="2026-03-13 09:37:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:37:37.405475619 +0000 UTC m=+1498.155390296" watchObservedRunningTime="2026-03-13 09:37:37.424578004 +0000 UTC m=+1498.174492681" Mar 13 09:37:37 crc kubenswrapper[4930]: I0313 09:37:37.671124 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8pbts"] Mar 13 09:37:38 crc kubenswrapper[4930]: I0313 09:37:38.397230 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8pbts" event={"ID":"7ccda94d-acb1-4c49-91e1-db24243cf6a9","Type":"ContainerStarted","Data":"e7f08920cc72e929f43e7095a0eee6181dd89a73297ca086d458bf87dfa2aa88"} Mar 13 09:37:38 crc kubenswrapper[4930]: I0313 09:37:38.397561 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8pbts" event={"ID":"7ccda94d-acb1-4c49-91e1-db24243cf6a9","Type":"ContainerStarted","Data":"e23894306634595e43297989684833d310dac36a4b3b425f22c6f4e0e7045249"} Mar 13 09:37:38 crc kubenswrapper[4930]: I0313 09:37:38.408163 4930 generic.go:334] "Generic (PLEG): container finished" podID="e5c00fcf-72f0-45e6-a1b3-235677ec4d14" containerID="e55d97b040c4f9dd7afc52670411587db48d28ec123410d9c0cd8f8198000fc7" exitCode=0 Mar 13 09:37:38 crc kubenswrapper[4930]: I0313 09:37:38.408269 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-3a0f-account-create-update-82l86" event={"ID":"e5c00fcf-72f0-45e6-a1b3-235677ec4d14","Type":"ContainerDied","Data":"e55d97b040c4f9dd7afc52670411587db48d28ec123410d9c0cd8f8198000fc7"} Mar 13 09:37:38 crc kubenswrapper[4930]: I0313 09:37:38.412861 4930 generic.go:334] "Generic (PLEG): container finished" podID="05b2d756-3fdc-4914-b511-350ee81fabb4" containerID="7c2d12ac50c8e12bd82ecd79c65662aa39860ec7c3c0512ad6f6716b34b3987d" exitCode=0 Mar 13 09:37:38 crc kubenswrapper[4930]: I0313 09:37:38.412898 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-r7nvk" event={"ID":"05b2d756-3fdc-4914-b511-350ee81fabb4","Type":"ContainerDied","Data":"7c2d12ac50c8e12bd82ecd79c65662aa39860ec7c3c0512ad6f6716b34b3987d"} Mar 13 09:37:38 crc kubenswrapper[4930]: I0313 09:37:38.415810 4930 generic.go:334] "Generic (PLEG): container finished" podID="ea17f053-2c9a-42ba-b729-106a9f9b90bb" containerID="bcf4931f3f7a5bb25aaa59f905b399eb539d28b86e0927e1bc0c7721e688be5b" exitCode=0 Mar 13 09:37:38 crc kubenswrapper[4930]: I0313 09:37:38.415852 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" event={"ID":"ea17f053-2c9a-42ba-b729-106a9f9b90bb","Type":"ContainerDied","Data":"bcf4931f3f7a5bb25aaa59f905b399eb539d28b86e0927e1bc0c7721e688be5b"} Mar 13 09:37:38 crc kubenswrapper[4930]: I0313 09:37:38.447658 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-8pbts" podStartSLOduration=2.447637208 podStartE2EDuration="2.447637208s" podCreationTimestamp="2026-03-13 09:37:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:37:38.428449821 +0000 UTC m=+1499.178364498" watchObservedRunningTime="2026-03-13 09:37:38.447637208 +0000 UTC m=+1499.197551905" Mar 13 09:37:38 crc kubenswrapper[4930]: I0313 09:37:38.499640 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:38 crc kubenswrapper[4930]: I0313 09:37:38.576493 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 09:37:39 crc kubenswrapper[4930]: I0313 09:37:39.263633 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:39 crc kubenswrapper[4930]: I0313 09:37:39.266238 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:37:39 crc kubenswrapper[4930]: I0313 09:37:39.453695 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" event={"ID":"ea17f053-2c9a-42ba-b729-106a9f9b90bb","Type":"ContainerStarted","Data":"0b3811dabc8c56096d68904a6457198a1363ac79593afecedd9a2b99a45876ce"} Mar 13 09:37:39 crc kubenswrapper[4930]: I0313 09:37:39.454521 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:39 crc kubenswrapper[4930]: I0313 09:37:39.490648 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" podStartSLOduration=5.490627807 podStartE2EDuration="5.490627807s" podCreationTimestamp="2026-03-13 09:37:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:37:39.48367515 +0000 UTC m=+1500.233589827" watchObservedRunningTime="2026-03-13 09:37:39.490627807 +0000 UTC m=+1500.240542484" Mar 13 09:37:40 crc kubenswrapper[4930]: E0313 09:37:40.000522 4930 info.go:109] Failed to get network devices: open /sys/class/net/7d0cca45d010cf2/address: no such file or directory Mar 13 09:37:40 crc kubenswrapper[4930]: I0313 09:37:40.336403 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-95hs2" podUID="0a86863c-63e1-4921-bd79-e6543a560043" containerName="registry-server" probeResult="failure" output=< Mar 13 09:37:40 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:37:40 crc kubenswrapper[4930]: > Mar 13 09:37:40 crc kubenswrapper[4930]: I0313 09:37:40.476558 4930 generic.go:334] "Generic (PLEG): container finished" podID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerID="1531e381efb54d1fc2d4bc432da7030534b007725ca21cf8e228d57f89d37dc5" exitCode=0 Mar 13 09:37:40 crc kubenswrapper[4930]: I0313 09:37:40.477601 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e40e80e4-7348-4e09-b661-c0c55c8c4a18","Type":"ContainerDied","Data":"1531e381efb54d1fc2d4bc432da7030534b007725ca21cf8e228d57f89d37dc5"} Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.720310 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-3a0f-account-create-update-82l86" Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.731767 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-r7nvk" Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.873932 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n7bz6\" (UniqueName: \"kubernetes.io/projected/05b2d756-3fdc-4914-b511-350ee81fabb4-kube-api-access-n7bz6\") pod \"05b2d756-3fdc-4914-b511-350ee81fabb4\" (UID: \"05b2d756-3fdc-4914-b511-350ee81fabb4\") " Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.874130 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5c00fcf-72f0-45e6-a1b3-235677ec4d14-operator-scripts\") pod \"e5c00fcf-72f0-45e6-a1b3-235677ec4d14\" (UID: \"e5c00fcf-72f0-45e6-a1b3-235677ec4d14\") " Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.874360 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05b2d756-3fdc-4914-b511-350ee81fabb4-operator-scripts\") pod \"05b2d756-3fdc-4914-b511-350ee81fabb4\" (UID: \"05b2d756-3fdc-4914-b511-350ee81fabb4\") " Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.874483 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mstdv\" (UniqueName: \"kubernetes.io/projected/e5c00fcf-72f0-45e6-a1b3-235677ec4d14-kube-api-access-mstdv\") pod \"e5c00fcf-72f0-45e6-a1b3-235677ec4d14\" (UID: \"e5c00fcf-72f0-45e6-a1b3-235677ec4d14\") " Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.875379 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5c00fcf-72f0-45e6-a1b3-235677ec4d14-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e5c00fcf-72f0-45e6-a1b3-235677ec4d14" (UID: "e5c00fcf-72f0-45e6-a1b3-235677ec4d14"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.876193 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05b2d756-3fdc-4914-b511-350ee81fabb4-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "05b2d756-3fdc-4914-b511-350ee81fabb4" (UID: "05b2d756-3fdc-4914-b511-350ee81fabb4"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.878052 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e5c00fcf-72f0-45e6-a1b3-235677ec4d14-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.878092 4930 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05b2d756-3fdc-4914-b511-350ee81fabb4-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.881511 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05b2d756-3fdc-4914-b511-350ee81fabb4-kube-api-access-n7bz6" (OuterVolumeSpecName: "kube-api-access-n7bz6") pod "05b2d756-3fdc-4914-b511-350ee81fabb4" (UID: "05b2d756-3fdc-4914-b511-350ee81fabb4"). InnerVolumeSpecName "kube-api-access-n7bz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.909107 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5c00fcf-72f0-45e6-a1b3-235677ec4d14-kube-api-access-mstdv" (OuterVolumeSpecName: "kube-api-access-mstdv") pod "e5c00fcf-72f0-45e6-a1b3-235677ec4d14" (UID: "e5c00fcf-72f0-45e6-a1b3-235677ec4d14"). InnerVolumeSpecName "kube-api-access-mstdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.982839 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mstdv\" (UniqueName: \"kubernetes.io/projected/e5c00fcf-72f0-45e6-a1b3-235677ec4d14-kube-api-access-mstdv\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:41 crc kubenswrapper[4930]: I0313 09:37:41.982873 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n7bz6\" (UniqueName: \"kubernetes.io/projected/05b2d756-3fdc-4914-b511-350ee81fabb4-kube-api-access-n7bz6\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.308443 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.308770 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.324410 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.389769 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-combined-ca-bundle\") pod \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.389819 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e40e80e4-7348-4e09-b661-c0c55c8c4a18-log-httpd\") pod \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.389868 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bwh7t\" (UniqueName: \"kubernetes.io/projected/e40e80e4-7348-4e09-b661-c0c55c8c4a18-kube-api-access-bwh7t\") pod \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.390078 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-config-data\") pod \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.390165 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-sg-core-conf-yaml\") pod \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.390224 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e40e80e4-7348-4e09-b661-c0c55c8c4a18-run-httpd\") pod \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.390285 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-scripts\") pod \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\" (UID: \"e40e80e4-7348-4e09-b661-c0c55c8c4a18\") " Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.390549 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e40e80e4-7348-4e09-b661-c0c55c8c4a18-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e40e80e4-7348-4e09-b661-c0c55c8c4a18" (UID: "e40e80e4-7348-4e09-b661-c0c55c8c4a18"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.391014 4930 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e40e80e4-7348-4e09-b661-c0c55c8c4a18-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.392557 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e40e80e4-7348-4e09-b661-c0c55c8c4a18-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e40e80e4-7348-4e09-b661-c0c55c8c4a18" (UID: "e40e80e4-7348-4e09-b661-c0c55c8c4a18"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.397850 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-scripts" (OuterVolumeSpecName: "scripts") pod "e40e80e4-7348-4e09-b661-c0c55c8c4a18" (UID: "e40e80e4-7348-4e09-b661-c0c55c8c4a18"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.409604 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e40e80e4-7348-4e09-b661-c0c55c8c4a18-kube-api-access-bwh7t" (OuterVolumeSpecName: "kube-api-access-bwh7t") pod "e40e80e4-7348-4e09-b661-c0c55c8c4a18" (UID: "e40e80e4-7348-4e09-b661-c0c55c8c4a18"). InnerVolumeSpecName "kube-api-access-bwh7t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.459589 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e40e80e4-7348-4e09-b661-c0c55c8c4a18" (UID: "e40e80e4-7348-4e09-b661-c0c55c8c4a18"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.492891 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bwh7t\" (UniqueName: \"kubernetes.io/projected/e40e80e4-7348-4e09-b661-c0c55c8c4a18-kube-api-access-bwh7t\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.492915 4930 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.492926 4930 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e40e80e4-7348-4e09-b661-c0c55c8c4a18-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.492936 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.520782 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-3a0f-account-create-update-82l86" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.521114 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-3a0f-account-create-update-82l86" event={"ID":"e5c00fcf-72f0-45e6-a1b3-235677ec4d14","Type":"ContainerDied","Data":"b0a53bef955941e2fce6f29f916d75e3e7df3db60a28b461597f5b2379af698c"} Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.521178 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0a53bef955941e2fce6f29f916d75e3e7df3db60a28b461597f5b2379af698c" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.530207 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e40e80e4-7348-4e09-b661-c0c55c8c4a18","Type":"ContainerDied","Data":"516962bc1d9e7b69a9366483f9ee70c6f5acd20b3b46c1c003409ebf2725cf7c"} Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.530262 4930 scope.go:117] "RemoveContainer" containerID="035f8969f11fbec2c070d2865cb2f2b53e02519f631b9e537036f3df8bbcc70c" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.530401 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.538466 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-r7nvk" event={"ID":"05b2d756-3fdc-4914-b511-350ee81fabb4","Type":"ContainerDied","Data":"7d0cca45d010cf2e7fbc97b2f83bbcb5f792119dfafc3348e727c37faca08a96"} Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.538501 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d0cca45d010cf2e7fbc97b2f83bbcb5f792119dfafc3348e727c37faca08a96" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.538579 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-r7nvk" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.575262 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e40e80e4-7348-4e09-b661-c0c55c8c4a18" (UID: "e40e80e4-7348-4e09-b661-c0c55c8c4a18"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.596159 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.631932 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-config-data" (OuterVolumeSpecName: "config-data") pod "e40e80e4-7348-4e09-b661-c0c55c8c4a18" (UID: "e40e80e4-7348-4e09-b661-c0c55c8c4a18"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.697323 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e40e80e4-7348-4e09-b661-c0c55c8c4a18-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.761778 4930 scope.go:117] "RemoveContainer" containerID="79f3124ee3849ab23fc945701f13ca3bee7f4016aa170a922bd6d7adc3d23179" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.838688 4930 scope.go:117] "RemoveContainer" containerID="fa7671ba97dda70ce688b6467bf102a454c07a2d07eb49426cd807b252b1c567" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.867671 4930 scope.go:117] "RemoveContainer" containerID="1531e381efb54d1fc2d4bc432da7030534b007725ca21cf8e228d57f89d37dc5" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.890565 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.914390 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.925779 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:42 crc kubenswrapper[4930]: E0313 09:37:42.926348 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5c00fcf-72f0-45e6-a1b3-235677ec4d14" containerName="mariadb-account-create-update" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.926379 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5c00fcf-72f0-45e6-a1b3-235677ec4d14" containerName="mariadb-account-create-update" Mar 13 09:37:42 crc kubenswrapper[4930]: E0313 09:37:42.926416 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="ceilometer-central-agent" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.926425 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="ceilometer-central-agent" Mar 13 09:37:42 crc kubenswrapper[4930]: E0313 09:37:42.926474 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="ceilometer-notification-agent" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.926484 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="ceilometer-notification-agent" Mar 13 09:37:42 crc kubenswrapper[4930]: E0313 09:37:42.926500 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="proxy-httpd" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.926508 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="proxy-httpd" Mar 13 09:37:42 crc kubenswrapper[4930]: E0313 09:37:42.926520 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="sg-core" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.926528 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="sg-core" Mar 13 09:37:42 crc kubenswrapper[4930]: E0313 09:37:42.926549 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05b2d756-3fdc-4914-b511-350ee81fabb4" containerName="mariadb-database-create" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.926557 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="05b2d756-3fdc-4914-b511-350ee81fabb4" containerName="mariadb-database-create" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.926817 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="proxy-httpd" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.926836 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="05b2d756-3fdc-4914-b511-350ee81fabb4" containerName="mariadb-database-create" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.926846 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="sg-core" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.926859 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5c00fcf-72f0-45e6-a1b3-235677ec4d14" containerName="mariadb-account-create-update" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.926880 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="ceilometer-central-agent" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.926899 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" containerName="ceilometer-notification-agent" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.929528 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.938593 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.960094 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:37:42 crc kubenswrapper[4930]: I0313 09:37:42.960283 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.112113 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-log-httpd\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.113238 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-config-data\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.113298 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-njls8\" (UniqueName: \"kubernetes.io/projected/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-kube-api-access-njls8\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.118757 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.118874 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-run-httpd\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.118968 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.118998 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-scripts\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.221311 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-config-data\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.221366 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-njls8\" (UniqueName: \"kubernetes.io/projected/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-kube-api-access-njls8\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.221405 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.221468 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-run-httpd\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.221510 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.221531 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-scripts\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.221656 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-log-httpd\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.222076 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-log-httpd\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.222267 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-run-httpd\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.225721 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.226335 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-config-data\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.227174 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.228577 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-scripts\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.241970 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-njls8\" (UniqueName: \"kubernetes.io/projected/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-kube-api-access-njls8\") pod \"ceilometer-0\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.292808 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.574910 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4b7ebf6b-9957-4833-be7b-2123a7a456ee","Type":"ContainerStarted","Data":"313f5fbaf4fd63adb12551298ae335950320ba02e956ae74e3d896436ab331d3"} Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.578698 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb","Type":"ContainerStarted","Data":"36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c"} Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.578857 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c" gracePeriod=30 Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.589373 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b1d06be6-d50e-4bc3-8579-7a97b48f32b2","Type":"ContainerStarted","Data":"8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6"} Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.589444 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b1d06be6-d50e-4bc3-8579-7a97b48f32b2","Type":"ContainerStarted","Data":"6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd"} Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.589598 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b1d06be6-d50e-4bc3-8579-7a97b48f32b2" containerName="nova-metadata-log" containerID="cri-o://6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd" gracePeriod=30 Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.589905 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b1d06be6-d50e-4bc3-8579-7a97b48f32b2" containerName="nova-metadata-metadata" containerID="cri-o://8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6" gracePeriod=30 Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.601879 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1a3fee40-f7de-42df-b078-abde4117d47a","Type":"ContainerStarted","Data":"0e4ee155f43fa38e80c115a947e61020cf8bfca7893d21cf41942b070572ec31"} Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.601920 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1a3fee40-f7de-42df-b078-abde4117d47a","Type":"ContainerStarted","Data":"eb7012e87694f097a1abb477294439e571a1ae1c1cc305ae5a03627d60319a7b"} Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.607352 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=4.164909148 podStartE2EDuration="9.607328604s" podCreationTimestamp="2026-03-13 09:37:34 +0000 UTC" firstStartedPulling="2026-03-13 09:37:36.973504941 +0000 UTC m=+1497.723419618" lastFinishedPulling="2026-03-13 09:37:42.415924407 +0000 UTC m=+1503.165839074" observedRunningTime="2026-03-13 09:37:43.591748008 +0000 UTC m=+1504.341662695" watchObservedRunningTime="2026-03-13 09:37:43.607328604 +0000 UTC m=+1504.357243291" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.635961 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.5837525230000002 podStartE2EDuration="9.63594303s" podCreationTimestamp="2026-03-13 09:37:34 +0000 UTC" firstStartedPulling="2026-03-13 09:37:36.220728079 +0000 UTC m=+1496.970642756" lastFinishedPulling="2026-03-13 09:37:42.272918576 +0000 UTC m=+1503.022833263" observedRunningTime="2026-03-13 09:37:43.615315146 +0000 UTC m=+1504.365229823" watchObservedRunningTime="2026-03-13 09:37:43.63594303 +0000 UTC m=+1504.385857707" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.649325 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.9819650429999998 podStartE2EDuration="9.649302879s" podCreationTimestamp="2026-03-13 09:37:34 +0000 UTC" firstStartedPulling="2026-03-13 09:37:36.602751028 +0000 UTC m=+1497.352665705" lastFinishedPulling="2026-03-13 09:37:42.270088864 +0000 UTC m=+1503.020003541" observedRunningTime="2026-03-13 09:37:43.639769587 +0000 UTC m=+1504.389684264" watchObservedRunningTime="2026-03-13 09:37:43.649302879 +0000 UTC m=+1504.399217556" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.679783 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.145010205 podStartE2EDuration="9.679760763s" podCreationTimestamp="2026-03-13 09:37:34 +0000 UTC" firstStartedPulling="2026-03-13 09:37:35.73391665 +0000 UTC m=+1496.483831327" lastFinishedPulling="2026-03-13 09:37:42.268667208 +0000 UTC m=+1503.018581885" observedRunningTime="2026-03-13 09:37:43.661577471 +0000 UTC m=+1504.411492158" watchObservedRunningTime="2026-03-13 09:37:43.679760763 +0000 UTC m=+1504.429675450" Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.877880 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:37:43 crc kubenswrapper[4930]: I0313 09:37:43.987117 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e40e80e4-7348-4e09-b661-c0c55c8c4a18" path="/var/lib/kubelet/pods/e40e80e4-7348-4e09-b661-c0c55c8c4a18/volumes" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.594168 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.615446 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce","Type":"ContainerStarted","Data":"0a4543fcaac00419d9510241502cf601d76933d8c901ac2c0b62dd3787af495a"} Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.622696 4930 generic.go:334] "Generic (PLEG): container finished" podID="b1d06be6-d50e-4bc3-8579-7a97b48f32b2" containerID="8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6" exitCode=0 Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.622729 4930 generic.go:334] "Generic (PLEG): container finished" podID="b1d06be6-d50e-4bc3-8579-7a97b48f32b2" containerID="6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd" exitCode=143 Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.623871 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.624368 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b1d06be6-d50e-4bc3-8579-7a97b48f32b2","Type":"ContainerDied","Data":"8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6"} Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.624400 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b1d06be6-d50e-4bc3-8579-7a97b48f32b2","Type":"ContainerDied","Data":"6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd"} Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.624412 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b1d06be6-d50e-4bc3-8579-7a97b48f32b2","Type":"ContainerDied","Data":"f208b782d2543d8e08baa20ea5215802363034935a2e9a8b6e5645efa1c6e0b2"} Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.624426 4930 scope.go:117] "RemoveContainer" containerID="8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.664180 4930 scope.go:117] "RemoveContainer" containerID="6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.691423 4930 scope.go:117] "RemoveContainer" containerID="8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6" Mar 13 09:37:44 crc kubenswrapper[4930]: E0313 09:37:44.693172 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6\": container with ID starting with 8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6 not found: ID does not exist" containerID="8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.693197 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6"} err="failed to get container status \"8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6\": rpc error: code = NotFound desc = could not find container \"8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6\": container with ID starting with 8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6 not found: ID does not exist" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.693219 4930 scope.go:117] "RemoveContainer" containerID="6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd" Mar 13 09:37:44 crc kubenswrapper[4930]: E0313 09:37:44.693666 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd\": container with ID starting with 6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd not found: ID does not exist" containerID="6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.693687 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd"} err="failed to get container status \"6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd\": rpc error: code = NotFound desc = could not find container \"6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd\": container with ID starting with 6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd not found: ID does not exist" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.693700 4930 scope.go:117] "RemoveContainer" containerID="8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.693917 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6"} err="failed to get container status \"8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6\": rpc error: code = NotFound desc = could not find container \"8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6\": container with ID starting with 8767e3d4eb00ca786a396fb85f6d85d54488fd5f1a4aa34e794ce27ca5f37bd6 not found: ID does not exist" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.693935 4930 scope.go:117] "RemoveContainer" containerID="6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.694135 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd"} err="failed to get container status \"6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd\": rpc error: code = NotFound desc = could not find container \"6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd\": container with ID starting with 6c502558d11c9d5e0d1a4853f46328340a7b234dc4690541b2e950963566dadd not found: ID does not exist" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.764526 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-config-data\") pod \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.764725 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92p2s\" (UniqueName: \"kubernetes.io/projected/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-kube-api-access-92p2s\") pod \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.764758 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-combined-ca-bundle\") pod \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.764832 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-logs\") pod \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\" (UID: \"b1d06be6-d50e-4bc3-8579-7a97b48f32b2\") " Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.765463 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-logs" (OuterVolumeSpecName: "logs") pod "b1d06be6-d50e-4bc3-8579-7a97b48f32b2" (UID: "b1d06be6-d50e-4bc3-8579-7a97b48f32b2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.765764 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.783450 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-kube-api-access-92p2s" (OuterVolumeSpecName: "kube-api-access-92p2s") pod "b1d06be6-d50e-4bc3-8579-7a97b48f32b2" (UID: "b1d06be6-d50e-4bc3-8579-7a97b48f32b2"). InnerVolumeSpecName "kube-api-access-92p2s". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.808676 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.808727 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.809796 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b1d06be6-d50e-4bc3-8579-7a97b48f32b2" (UID: "b1d06be6-d50e-4bc3-8579-7a97b48f32b2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.810976 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-config-data" (OuterVolumeSpecName: "config-data") pod "b1d06be6-d50e-4bc3-8579-7a97b48f32b2" (UID: "b1d06be6-d50e-4bc3-8579-7a97b48f32b2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.867981 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.868015 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92p2s\" (UniqueName: \"kubernetes.io/projected/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-kube-api-access-92p2s\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.868027 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b1d06be6-d50e-4bc3-8579-7a97b48f32b2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.963158 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:44 crc kubenswrapper[4930]: I0313 09:37:44.979912 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.000889 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:45 crc kubenswrapper[4930]: E0313 09:37:45.001380 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d06be6-d50e-4bc3-8579-7a97b48f32b2" containerName="nova-metadata-metadata" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.001393 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d06be6-d50e-4bc3-8579-7a97b48f32b2" containerName="nova-metadata-metadata" Mar 13 09:37:45 crc kubenswrapper[4930]: E0313 09:37:45.001420 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1d06be6-d50e-4bc3-8579-7a97b48f32b2" containerName="nova-metadata-log" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.001425 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1d06be6-d50e-4bc3-8579-7a97b48f32b2" containerName="nova-metadata-log" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.001741 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1d06be6-d50e-4bc3-8579-7a97b48f32b2" containerName="nova-metadata-metadata" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.001775 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1d06be6-d50e-4bc3-8579-7a97b48f32b2" containerName="nova-metadata-log" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.003007 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.008785 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.008953 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.036228 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.135629 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.146484 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-dh2pc"] Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.148115 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.152766 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.153035 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.153365 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.153592 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dc4s9" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.160399 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-dh2pc"] Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.175964 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-config-data\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.176156 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.176248 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ftnq9\" (UniqueName: \"kubernetes.io/projected/68f691c8-ea38-424b-b373-1afbfdf9df4a-kube-api-access-ftnq9\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.176341 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f691c8-ea38-424b-b373-1afbfdf9df4a-logs\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.176646 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.278976 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-config-data\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.279062 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-combined-ca-bundle\") pod \"aodh-db-sync-dh2pc\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.279087 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.279109 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ftnq9\" (UniqueName: \"kubernetes.io/projected/68f691c8-ea38-424b-b373-1afbfdf9df4a-kube-api-access-ftnq9\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.279182 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f691c8-ea38-424b-b373-1afbfdf9df4a-logs\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.279221 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzvbw\" (UniqueName: \"kubernetes.io/projected/17fdbf0c-1335-4369-84b3-7db681759d19-kube-api-access-nzvbw\") pod \"aodh-db-sync-dh2pc\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.279247 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.279312 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-config-data\") pod \"aodh-db-sync-dh2pc\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.279335 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-scripts\") pod \"aodh-db-sync-dh2pc\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.280141 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f691c8-ea38-424b-b373-1afbfdf9df4a-logs\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.286378 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.288117 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-config-data\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.293222 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.302056 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ftnq9\" (UniqueName: \"kubernetes.io/projected/68f691c8-ea38-424b-b373-1afbfdf9df4a-kube-api-access-ftnq9\") pod \"nova-metadata-0\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.326063 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.326745 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.326863 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.362701 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.381788 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzvbw\" (UniqueName: \"kubernetes.io/projected/17fdbf0c-1335-4369-84b3-7db681759d19-kube-api-access-nzvbw\") pod \"aodh-db-sync-dh2pc\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.382096 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-config-data\") pod \"aodh-db-sync-dh2pc\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.382204 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-scripts\") pod \"aodh-db-sync-dh2pc\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.382391 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-combined-ca-bundle\") pod \"aodh-db-sync-dh2pc\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.386329 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-combined-ca-bundle\") pod \"aodh-db-sync-dh2pc\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.388357 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-config-data\") pod \"aodh-db-sync-dh2pc\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.391852 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-scripts\") pod \"aodh-db-sync-dh2pc\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.399114 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.417113 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzvbw\" (UniqueName: \"kubernetes.io/projected/17fdbf0c-1335-4369-84b3-7db681759d19-kube-api-access-nzvbw\") pod \"aodh-db-sync-dh2pc\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.452926 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-hr7c4"] Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.453221 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" podUID="a2e994f9-023c-4cb7-b6b7-5a974a78e746" containerName="dnsmasq-dns" containerID="cri-o://76e3a37a915bf1d9b57c8869108d42f44b5f2a831bfa3cef11772ed1074b344f" gracePeriod=10 Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.499924 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.713944 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce","Type":"ContainerStarted","Data":"aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77"} Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.714272 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce","Type":"ContainerStarted","Data":"93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36"} Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.719145 4930 generic.go:334] "Generic (PLEG): container finished" podID="a2e994f9-023c-4cb7-b6b7-5a974a78e746" containerID="76e3a37a915bf1d9b57c8869108d42f44b5f2a831bfa3cef11772ed1074b344f" exitCode=0 Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.720273 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" event={"ID":"a2e994f9-023c-4cb7-b6b7-5a974a78e746","Type":"ContainerDied","Data":"76e3a37a915bf1d9b57c8869108d42f44b5f2a831bfa3cef11772ed1074b344f"} Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.789632 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.911588 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1a3fee40-f7de-42df-b078-abde4117d47a" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.250:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 09:37:45 crc kubenswrapper[4930]: I0313 09:37:45.911951 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="1a3fee40-f7de-42df-b078-abde4117d47a" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.250:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.010021 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1d06be6-d50e-4bc3-8579-7a97b48f32b2" path="/var/lib/kubelet/pods/b1d06be6-d50e-4bc3-8579-7a97b48f32b2/volumes" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.059896 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.277467 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-dh2pc"] Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.451602 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.629044 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-dns-svc\") pod \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.629482 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-config\") pod \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.629892 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-ovsdbserver-sb\") pod \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.630023 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-dns-swift-storage-0\") pod \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.630219 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-ovsdbserver-nb\") pod \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.630298 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqf7d\" (UniqueName: \"kubernetes.io/projected/a2e994f9-023c-4cb7-b6b7-5a974a78e746-kube-api-access-sqf7d\") pod \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\" (UID: \"a2e994f9-023c-4cb7-b6b7-5a974a78e746\") " Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.678332 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2e994f9-023c-4cb7-b6b7-5a974a78e746-kube-api-access-sqf7d" (OuterVolumeSpecName: "kube-api-access-sqf7d") pod "a2e994f9-023c-4cb7-b6b7-5a974a78e746" (UID: "a2e994f9-023c-4cb7-b6b7-5a974a78e746"). InnerVolumeSpecName "kube-api-access-sqf7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.718476 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-config" (OuterVolumeSpecName: "config") pod "a2e994f9-023c-4cb7-b6b7-5a974a78e746" (UID: "a2e994f9-023c-4cb7-b6b7-5a974a78e746"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.735056 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.735092 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqf7d\" (UniqueName: \"kubernetes.io/projected/a2e994f9-023c-4cb7-b6b7-5a974a78e746-kube-api-access-sqf7d\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.810640 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a2e994f9-023c-4cb7-b6b7-5a974a78e746" (UID: "a2e994f9-023c-4cb7-b6b7-5a974a78e746"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.829901 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "a2e994f9-023c-4cb7-b6b7-5a974a78e746" (UID: "a2e994f9-023c-4cb7-b6b7-5a974a78e746"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.832931 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce","Type":"ContainerStarted","Data":"716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df"} Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.838466 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.838495 4930 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.877747 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" event={"ID":"a2e994f9-023c-4cb7-b6b7-5a974a78e746","Type":"ContainerDied","Data":"ae7e4431da1cc8392755493e685b1b9a59e9329e5af0b0c4b3361f20a3042872"} Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.877796 4930 scope.go:117] "RemoveContainer" containerID="76e3a37a915bf1d9b57c8869108d42f44b5f2a831bfa3cef11772ed1074b344f" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.877940 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-688b9f5b49-hr7c4" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.887784 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a2e994f9-023c-4cb7-b6b7-5a974a78e746" (UID: "a2e994f9-023c-4cb7-b6b7-5a974a78e746"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.898186 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-dh2pc" event={"ID":"17fdbf0c-1335-4369-84b3-7db681759d19","Type":"ContainerStarted","Data":"84e981fbaa90967b8663bf2b9952c73c82fd4c03b24f07b79f3a25356e817ec7"} Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.914826 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68f691c8-ea38-424b-b373-1afbfdf9df4a","Type":"ContainerStarted","Data":"1510f9469e8dabcb063ab8132ce15bc8f57f6b97f7da372ead4777cb65bf56b9"} Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.914871 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68f691c8-ea38-424b-b373-1afbfdf9df4a","Type":"ContainerStarted","Data":"355ccac42b26a18749157862ed4e424678d9911f8192a578814d7d3f4fc2d93f"} Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.925979 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a2e994f9-023c-4cb7-b6b7-5a974a78e746" (UID: "a2e994f9-023c-4cb7-b6b7-5a974a78e746"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.940210 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.940320 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a2e994f9-023c-4cb7-b6b7-5a974a78e746-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:46 crc kubenswrapper[4930]: I0313 09:37:46.941726 4930 scope.go:117] "RemoveContainer" containerID="359ba1e318d6db010d5d820cadf70025e50d9954b26f32eeaa8ccf66e62215f2" Mar 13 09:37:47 crc kubenswrapper[4930]: I0313 09:37:47.241492 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-hr7c4"] Mar 13 09:37:47 crc kubenswrapper[4930]: I0313 09:37:47.248192 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-688b9f5b49-hr7c4"] Mar 13 09:37:47 crc kubenswrapper[4930]: I0313 09:37:47.924947 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68f691c8-ea38-424b-b373-1afbfdf9df4a","Type":"ContainerStarted","Data":"7b85c110a25ef12d1b7bb9c7e894f32dbd9d14e222c96200d1fd8a2ffa52d09e"} Mar 13 09:37:47 crc kubenswrapper[4930]: I0313 09:37:47.927889 4930 generic.go:334] "Generic (PLEG): container finished" podID="67ceb05f-14a6-4adb-af90-5d52549e4f23" containerID="3b32b5ccd512cccbb3e04056bfad8de1f17b714917b17aac986da63abd41b2f0" exitCode=0 Mar 13 09:37:47 crc kubenswrapper[4930]: I0313 09:37:47.927947 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hljq9" event={"ID":"67ceb05f-14a6-4adb-af90-5d52549e4f23","Type":"ContainerDied","Data":"3b32b5ccd512cccbb3e04056bfad8de1f17b714917b17aac986da63abd41b2f0"} Mar 13 09:37:47 crc kubenswrapper[4930]: I0313 09:37:47.947277 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.947258018 podStartE2EDuration="3.947258018s" podCreationTimestamp="2026-03-13 09:37:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:37:47.944348974 +0000 UTC m=+1508.694263651" watchObservedRunningTime="2026-03-13 09:37:47.947258018 +0000 UTC m=+1508.697172695" Mar 13 09:37:47 crc kubenswrapper[4930]: I0313 09:37:47.984383 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2e994f9-023c-4cb7-b6b7-5a974a78e746" path="/var/lib/kubelet/pods/a2e994f9-023c-4cb7-b6b7-5a974a78e746/volumes" Mar 13 09:37:48 crc kubenswrapper[4930]: I0313 09:37:48.949044 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce","Type":"ContainerStarted","Data":"dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359"} Mar 13 09:37:48 crc kubenswrapper[4930]: I0313 09:37:48.949478 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 09:37:48 crc kubenswrapper[4930]: I0313 09:37:48.991624 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.293747401 podStartE2EDuration="6.991597863s" podCreationTimestamp="2026-03-13 09:37:42 +0000 UTC" firstStartedPulling="2026-03-13 09:37:43.878648292 +0000 UTC m=+1504.628562969" lastFinishedPulling="2026-03-13 09:37:48.576498754 +0000 UTC m=+1509.326413431" observedRunningTime="2026-03-13 09:37:48.977915635 +0000 UTC m=+1509.727830342" watchObservedRunningTime="2026-03-13 09:37:48.991597863 +0000 UTC m=+1509.741512540" Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.520855 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.649565 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-combined-ca-bundle\") pod \"67ceb05f-14a6-4adb-af90-5d52549e4f23\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.649707 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-scripts\") pod \"67ceb05f-14a6-4adb-af90-5d52549e4f23\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.649850 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-config-data\") pod \"67ceb05f-14a6-4adb-af90-5d52549e4f23\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.649957 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2zs2\" (UniqueName: \"kubernetes.io/projected/67ceb05f-14a6-4adb-af90-5d52549e4f23-kube-api-access-j2zs2\") pod \"67ceb05f-14a6-4adb-af90-5d52549e4f23\" (UID: \"67ceb05f-14a6-4adb-af90-5d52549e4f23\") " Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.668165 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-scripts" (OuterVolumeSpecName: "scripts") pod "67ceb05f-14a6-4adb-af90-5d52549e4f23" (UID: "67ceb05f-14a6-4adb-af90-5d52549e4f23"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.674876 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67ceb05f-14a6-4adb-af90-5d52549e4f23-kube-api-access-j2zs2" (OuterVolumeSpecName: "kube-api-access-j2zs2") pod "67ceb05f-14a6-4adb-af90-5d52549e4f23" (UID: "67ceb05f-14a6-4adb-af90-5d52549e4f23"). InnerVolumeSpecName "kube-api-access-j2zs2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.697998 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-config-data" (OuterVolumeSpecName: "config-data") pod "67ceb05f-14a6-4adb-af90-5d52549e4f23" (UID: "67ceb05f-14a6-4adb-af90-5d52549e4f23"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.702385 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "67ceb05f-14a6-4adb-af90-5d52549e4f23" (UID: "67ceb05f-14a6-4adb-af90-5d52549e4f23"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.755061 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2zs2\" (UniqueName: \"kubernetes.io/projected/67ceb05f-14a6-4adb-af90-5d52549e4f23-kube-api-access-j2zs2\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.755646 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.755658 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.755670 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67ceb05f-14a6-4adb-af90-5d52549e4f23-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.966176 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-hljq9" Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.966170 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-hljq9" event={"ID":"67ceb05f-14a6-4adb-af90-5d52549e4f23","Type":"ContainerDied","Data":"0af727263de5b4a55896b7077a62e63e28522d786577b7eb02b7df3c951b3467"} Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.966231 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0af727263de5b4a55896b7077a62e63e28522d786577b7eb02b7df3c951b3467" Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.968870 4930 generic.go:334] "Generic (PLEG): container finished" podID="7ccda94d-acb1-4c49-91e1-db24243cf6a9" containerID="e7f08920cc72e929f43e7095a0eee6181dd89a73297ca086d458bf87dfa2aa88" exitCode=0 Mar 13 09:37:49 crc kubenswrapper[4930]: I0313 09:37:49.970243 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8pbts" event={"ID":"7ccda94d-acb1-4c49-91e1-db24243cf6a9","Type":"ContainerDied","Data":"e7f08920cc72e929f43e7095a0eee6181dd89a73297ca086d458bf87dfa2aa88"} Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.198961 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.199230 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1a3fee40-f7de-42df-b078-abde4117d47a" containerName="nova-api-log" containerID="cri-o://eb7012e87694f097a1abb477294439e571a1ae1c1cc305ae5a03627d60319a7b" gracePeriod=30 Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.199664 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="1a3fee40-f7de-42df-b078-abde4117d47a" containerName="nova-api-api" containerID="cri-o://0e4ee155f43fa38e80c115a947e61020cf8bfca7893d21cf41942b070572ec31" gracePeriod=30 Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.226581 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.226885 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="4b7ebf6b-9957-4833-be7b-2123a7a456ee" containerName="nova-scheduler-scheduler" containerID="cri-o://313f5fbaf4fd63adb12551298ae335950320ba02e956ae74e3d896436ab331d3" gracePeriod=30 Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.240609 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.240851 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="68f691c8-ea38-424b-b373-1afbfdf9df4a" containerName="nova-metadata-log" containerID="cri-o://1510f9469e8dabcb063ab8132ce15bc8f57f6b97f7da372ead4777cb65bf56b9" gracePeriod=30 Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.240999 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="68f691c8-ea38-424b-b373-1afbfdf9df4a" containerName="nova-metadata-metadata" containerID="cri-o://7b85c110a25ef12d1b7bb9c7e894f32dbd9d14e222c96200d1fd8a2ffa52d09e" gracePeriod=30 Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.327575 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.327623 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 09:37:50 crc kubenswrapper[4930]: E0313 09:37:50.329841 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="313f5fbaf4fd63adb12551298ae335950320ba02e956ae74e3d896436ab331d3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 09:37:50 crc kubenswrapper[4930]: E0313 09:37:50.331414 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="313f5fbaf4fd63adb12551298ae335950320ba02e956ae74e3d896436ab331d3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 09:37:50 crc kubenswrapper[4930]: E0313 09:37:50.336225 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="313f5fbaf4fd63adb12551298ae335950320ba02e956ae74e3d896436ab331d3" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 09:37:50 crc kubenswrapper[4930]: E0313 09:37:50.336283 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="4b7ebf6b-9957-4833-be7b-2123a7a456ee" containerName="nova-scheduler-scheduler" Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.372627 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-95hs2" podUID="0a86863c-63e1-4921-bd79-e6543a560043" containerName="registry-server" probeResult="failure" output=< Mar 13 09:37:50 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:37:50 crc kubenswrapper[4930]: > Mar 13 09:37:50 crc kubenswrapper[4930]: E0313 09:37:50.379976 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68f691c8_ea38_424b_b373_1afbfdf9df4a.slice/crio-1510f9469e8dabcb063ab8132ce15bc8f57f6b97f7da372ead4777cb65bf56b9.scope\": RecentStats: unable to find data in memory cache]" Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.993677 4930 generic.go:334] "Generic (PLEG): container finished" podID="68f691c8-ea38-424b-b373-1afbfdf9df4a" containerID="7b85c110a25ef12d1b7bb9c7e894f32dbd9d14e222c96200d1fd8a2ffa52d09e" exitCode=0 Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.994000 4930 generic.go:334] "Generic (PLEG): container finished" podID="68f691c8-ea38-424b-b373-1afbfdf9df4a" containerID="1510f9469e8dabcb063ab8132ce15bc8f57f6b97f7da372ead4777cb65bf56b9" exitCode=143 Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.994070 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68f691c8-ea38-424b-b373-1afbfdf9df4a","Type":"ContainerDied","Data":"7b85c110a25ef12d1b7bb9c7e894f32dbd9d14e222c96200d1fd8a2ffa52d09e"} Mar 13 09:37:50 crc kubenswrapper[4930]: I0313 09:37:50.994099 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68f691c8-ea38-424b-b373-1afbfdf9df4a","Type":"ContainerDied","Data":"1510f9469e8dabcb063ab8132ce15bc8f57f6b97f7da372ead4777cb65bf56b9"} Mar 13 09:37:51 crc kubenswrapper[4930]: I0313 09:37:51.006365 4930 generic.go:334] "Generic (PLEG): container finished" podID="1a3fee40-f7de-42df-b078-abde4117d47a" containerID="eb7012e87694f097a1abb477294439e571a1ae1c1cc305ae5a03627d60319a7b" exitCode=143 Mar 13 09:37:51 crc kubenswrapper[4930]: I0313 09:37:51.006483 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1a3fee40-f7de-42df-b078-abde4117d47a","Type":"ContainerDied","Data":"eb7012e87694f097a1abb477294439e571a1ae1c1cc305ae5a03627d60319a7b"} Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.055198 4930 generic.go:334] "Generic (PLEG): container finished" podID="4b7ebf6b-9957-4833-be7b-2123a7a456ee" containerID="313f5fbaf4fd63adb12551298ae335950320ba02e956ae74e3d896436ab331d3" exitCode=0 Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.055291 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4b7ebf6b-9957-4833-be7b-2123a7a456ee","Type":"ContainerDied","Data":"313f5fbaf4fd63adb12551298ae335950320ba02e956ae74e3d896436ab331d3"} Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.272281 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.359755 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-config-data\") pod \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.359943 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxn67\" (UniqueName: \"kubernetes.io/projected/7ccda94d-acb1-4c49-91e1-db24243cf6a9-kube-api-access-qxn67\") pod \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.360024 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-scripts\") pod \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.360293 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-combined-ca-bundle\") pod \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\" (UID: \"7ccda94d-acb1-4c49-91e1-db24243cf6a9\") " Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.369649 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-scripts" (OuterVolumeSpecName: "scripts") pod "7ccda94d-acb1-4c49-91e1-db24243cf6a9" (UID: "7ccda94d-acb1-4c49-91e1-db24243cf6a9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.369760 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7ccda94d-acb1-4c49-91e1-db24243cf6a9-kube-api-access-qxn67" (OuterVolumeSpecName: "kube-api-access-qxn67") pod "7ccda94d-acb1-4c49-91e1-db24243cf6a9" (UID: "7ccda94d-acb1-4c49-91e1-db24243cf6a9"). InnerVolumeSpecName "kube-api-access-qxn67". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.393465 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7ccda94d-acb1-4c49-91e1-db24243cf6a9" (UID: "7ccda94d-acb1-4c49-91e1-db24243cf6a9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.466096 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.466330 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.466397 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxn67\" (UniqueName: \"kubernetes.io/projected/7ccda94d-acb1-4c49-91e1-db24243cf6a9-kube-api-access-qxn67\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.570416 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-config-data" (OuterVolumeSpecName: "config-data") pod "7ccda94d-acb1-4c49-91e1-db24243cf6a9" (UID: "7ccda94d-acb1-4c49-91e1-db24243cf6a9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.678569 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7ccda94d-acb1-4c49-91e1-db24243cf6a9-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.810765 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.884279 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ftnq9\" (UniqueName: \"kubernetes.io/projected/68f691c8-ea38-424b-b373-1afbfdf9df4a-kube-api-access-ftnq9\") pod \"68f691c8-ea38-424b-b373-1afbfdf9df4a\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.884451 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-nova-metadata-tls-certs\") pod \"68f691c8-ea38-424b-b373-1afbfdf9df4a\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.884481 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f691c8-ea38-424b-b373-1afbfdf9df4a-logs\") pod \"68f691c8-ea38-424b-b373-1afbfdf9df4a\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.884619 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-combined-ca-bundle\") pod \"68f691c8-ea38-424b-b373-1afbfdf9df4a\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.884701 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-config-data\") pod \"68f691c8-ea38-424b-b373-1afbfdf9df4a\" (UID: \"68f691c8-ea38-424b-b373-1afbfdf9df4a\") " Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.885151 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68f691c8-ea38-424b-b373-1afbfdf9df4a-logs" (OuterVolumeSpecName: "logs") pod "68f691c8-ea38-424b-b373-1afbfdf9df4a" (UID: "68f691c8-ea38-424b-b373-1afbfdf9df4a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.885855 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/68f691c8-ea38-424b-b373-1afbfdf9df4a-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.900143 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68f691c8-ea38-424b-b373-1afbfdf9df4a-kube-api-access-ftnq9" (OuterVolumeSpecName: "kube-api-access-ftnq9") pod "68f691c8-ea38-424b-b373-1afbfdf9df4a" (UID: "68f691c8-ea38-424b-b373-1afbfdf9df4a"). InnerVolumeSpecName "kube-api-access-ftnq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.931329 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68f691c8-ea38-424b-b373-1afbfdf9df4a" (UID: "68f691c8-ea38-424b-b373-1afbfdf9df4a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.960640 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-config-data" (OuterVolumeSpecName: "config-data") pod "68f691c8-ea38-424b-b373-1afbfdf9df4a" (UID: "68f691c8-ea38-424b-b373-1afbfdf9df4a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.988614 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ftnq9\" (UniqueName: \"kubernetes.io/projected/68f691c8-ea38-424b-b373-1afbfdf9df4a-kube-api-access-ftnq9\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.988957 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:53 crc kubenswrapper[4930]: I0313 09:37:53.988972 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.062266 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "68f691c8-ea38-424b-b373-1afbfdf9df4a" (UID: "68f691c8-ea38-424b-b373-1afbfdf9df4a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.073845 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-8pbts" event={"ID":"7ccda94d-acb1-4c49-91e1-db24243cf6a9","Type":"ContainerDied","Data":"e23894306634595e43297989684833d310dac36a4b3b425f22c6f4e0e7045249"} Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.073893 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e23894306634595e43297989684833d310dac36a4b3b425f22c6f4e0e7045249" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.073961 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-8pbts" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.076870 4930 generic.go:334] "Generic (PLEG): container finished" podID="1a3fee40-f7de-42df-b078-abde4117d47a" containerID="0e4ee155f43fa38e80c115a947e61020cf8bfca7893d21cf41942b070572ec31" exitCode=0 Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.076937 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1a3fee40-f7de-42df-b078-abde4117d47a","Type":"ContainerDied","Data":"0e4ee155f43fa38e80c115a947e61020cf8bfca7893d21cf41942b070572ec31"} Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.078559 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"68f691c8-ea38-424b-b373-1afbfdf9df4a","Type":"ContainerDied","Data":"355ccac42b26a18749157862ed4e424678d9911f8192a578814d7d3f4fc2d93f"} Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.078591 4930 scope.go:117] "RemoveContainer" containerID="7b85c110a25ef12d1b7bb9c7e894f32dbd9d14e222c96200d1fd8a2ffa52d09e" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.078637 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.080231 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-dh2pc" event={"ID":"17fdbf0c-1335-4369-84b3-7db681759d19","Type":"ContainerStarted","Data":"08020393b80436228d130190c9583a49ddda5ae786b94588b37ff58a1ad73db6"} Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.091761 4930 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/68f691c8-ea38-424b-b373-1afbfdf9df4a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.121257 4930 scope.go:117] "RemoveContainer" containerID="1510f9469e8dabcb063ab8132ce15bc8f57f6b97f7da372ead4777cb65bf56b9" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.121460 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-dh2pc" podStartSLOduration=1.7193723539999999 podStartE2EDuration="9.1214191s" podCreationTimestamp="2026-03-13 09:37:45 +0000 UTC" firstStartedPulling="2026-03-13 09:37:46.298715284 +0000 UTC m=+1507.048629961" lastFinishedPulling="2026-03-13 09:37:53.70076203 +0000 UTC m=+1514.450676707" observedRunningTime="2026-03-13 09:37:54.094328412 +0000 UTC m=+1514.844243089" watchObservedRunningTime="2026-03-13 09:37:54.1214191 +0000 UTC m=+1514.871333777" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.171595 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.185765 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.200185 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:54 crc kubenswrapper[4930]: E0313 09:37:54.201667 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e994f9-023c-4cb7-b6b7-5a974a78e746" containerName="init" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.201687 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e994f9-023c-4cb7-b6b7-5a974a78e746" containerName="init" Mar 13 09:37:54 crc kubenswrapper[4930]: E0313 09:37:54.201699 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7ccda94d-acb1-4c49-91e1-db24243cf6a9" containerName="nova-cell1-conductor-db-sync" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.201705 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="7ccda94d-acb1-4c49-91e1-db24243cf6a9" containerName="nova-cell1-conductor-db-sync" Mar 13 09:37:54 crc kubenswrapper[4930]: E0313 09:37:54.201720 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2e994f9-023c-4cb7-b6b7-5a974a78e746" containerName="dnsmasq-dns" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.201726 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2e994f9-023c-4cb7-b6b7-5a974a78e746" containerName="dnsmasq-dns" Mar 13 09:37:54 crc kubenswrapper[4930]: E0313 09:37:54.201747 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ceb05f-14a6-4adb-af90-5d52549e4f23" containerName="nova-manage" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.201753 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ceb05f-14a6-4adb-af90-5d52549e4f23" containerName="nova-manage" Mar 13 09:37:54 crc kubenswrapper[4930]: E0313 09:37:54.201780 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f691c8-ea38-424b-b373-1afbfdf9df4a" containerName="nova-metadata-log" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.201787 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f691c8-ea38-424b-b373-1afbfdf9df4a" containerName="nova-metadata-log" Mar 13 09:37:54 crc kubenswrapper[4930]: E0313 09:37:54.201799 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f691c8-ea38-424b-b373-1afbfdf9df4a" containerName="nova-metadata-metadata" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.201805 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f691c8-ea38-424b-b373-1afbfdf9df4a" containerName="nova-metadata-metadata" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.201999 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2e994f9-023c-4cb7-b6b7-5a974a78e746" containerName="dnsmasq-dns" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.202019 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="7ccda94d-acb1-4c49-91e1-db24243cf6a9" containerName="nova-cell1-conductor-db-sync" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.202042 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="68f691c8-ea38-424b-b373-1afbfdf9df4a" containerName="nova-metadata-log" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.202053 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="68f691c8-ea38-424b-b373-1afbfdf9df4a" containerName="nova-metadata-metadata" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.202067 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="67ceb05f-14a6-4adb-af90-5d52549e4f23" containerName="nova-manage" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.207700 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.211524 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.212136 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.216307 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.357181 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2crfk\" (UniqueName: \"kubernetes.io/projected/b45be8d9-5062-4d02-8a1b-3597f9c40c60-kube-api-access-2crfk\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.357453 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.358630 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.362571 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-config-data\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.362798 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.363057 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b45be8d9-5062-4d02-8a1b-3597f9c40c60-logs\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.375624 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 09:37:54 crc kubenswrapper[4930]: E0313 09:37:54.418211 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b7ebf6b-9957-4833-be7b-2123a7a456ee" containerName="nova-scheduler-scheduler" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.418412 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b7ebf6b-9957-4833-be7b-2123a7a456ee" containerName="nova-scheduler-scheduler" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.419619 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b7ebf6b-9957-4833-be7b-2123a7a456ee" containerName="nova-scheduler-scheduler" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.424991 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.432325 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.437710 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.457135 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.492561 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvctp\" (UniqueName: \"kubernetes.io/projected/4b7ebf6b-9957-4833-be7b-2123a7a456ee-kube-api-access-vvctp\") pod \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\" (UID: \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\") " Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.492707 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b7ebf6b-9957-4833-be7b-2123a7a456ee-config-data\") pod \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\" (UID: \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\") " Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.492799 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7ebf6b-9957-4833-be7b-2123a7a456ee-combined-ca-bundle\") pod \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\" (UID: \"4b7ebf6b-9957-4833-be7b-2123a7a456ee\") " Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.493376 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.493543 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-config-data\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.493615 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.494195 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b45be8d9-5062-4d02-8a1b-3597f9c40c60-logs\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.494369 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2crfk\" (UniqueName: \"kubernetes.io/projected/b45be8d9-5062-4d02-8a1b-3597f9c40c60-kube-api-access-2crfk\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.495230 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b45be8d9-5062-4d02-8a1b-3597f9c40c60-logs\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.497742 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b7ebf6b-9957-4833-be7b-2123a7a456ee-kube-api-access-vvctp" (OuterVolumeSpecName: "kube-api-access-vvctp") pod "4b7ebf6b-9957-4833-be7b-2123a7a456ee" (UID: "4b7ebf6b-9957-4833-be7b-2123a7a456ee"). InnerVolumeSpecName "kube-api-access-vvctp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.500351 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-config-data\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.501001 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.508453 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.514043 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2crfk\" (UniqueName: \"kubernetes.io/projected/b45be8d9-5062-4d02-8a1b-3597f9c40c60-kube-api-access-2crfk\") pod \"nova-metadata-0\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.543993 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b7ebf6b-9957-4833-be7b-2123a7a456ee-config-data" (OuterVolumeSpecName: "config-data") pod "4b7ebf6b-9957-4833-be7b-2123a7a456ee" (UID: "4b7ebf6b-9957-4833-be7b-2123a7a456ee"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.554557 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b7ebf6b-9957-4833-be7b-2123a7a456ee-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b7ebf6b-9957-4833-be7b-2123a7a456ee" (UID: "4b7ebf6b-9957-4833-be7b-2123a7a456ee"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.596419 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xljc6\" (UniqueName: \"kubernetes.io/projected/1a3fee40-f7de-42df-b078-abde4117d47a-kube-api-access-xljc6\") pod \"1a3fee40-f7de-42df-b078-abde4117d47a\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.596672 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3fee40-f7de-42df-b078-abde4117d47a-combined-ca-bundle\") pod \"1a3fee40-f7de-42df-b078-abde4117d47a\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.596895 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a3fee40-f7de-42df-b078-abde4117d47a-logs\") pod \"1a3fee40-f7de-42df-b078-abde4117d47a\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.596966 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3fee40-f7de-42df-b078-abde4117d47a-config-data\") pod \"1a3fee40-f7de-42df-b078-abde4117d47a\" (UID: \"1a3fee40-f7de-42df-b078-abde4117d47a\") " Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.597213 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a3fee40-f7de-42df-b078-abde4117d47a-logs" (OuterVolumeSpecName: "logs") pod "1a3fee40-f7de-42df-b078-abde4117d47a" (UID: "1a3fee40-f7de-42df-b078-abde4117d47a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.597559 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r96p\" (UniqueName: \"kubernetes.io/projected/729e49b1-0bf4-43cf-b184-973dda23dc9c-kube-api-access-8r96p\") pod \"nova-cell1-conductor-0\" (UID: \"729e49b1-0bf4-43cf-b184-973dda23dc9c\") " pod="openstack/nova-cell1-conductor-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.597777 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729e49b1-0bf4-43cf-b184-973dda23dc9c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"729e49b1-0bf4-43cf-b184-973dda23dc9c\") " pod="openstack/nova-cell1-conductor-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.597879 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729e49b1-0bf4-43cf-b184-973dda23dc9c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"729e49b1-0bf4-43cf-b184-973dda23dc9c\") " pod="openstack/nova-cell1-conductor-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.598096 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b7ebf6b-9957-4833-be7b-2123a7a456ee-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.598112 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b7ebf6b-9957-4833-be7b-2123a7a456ee-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.598126 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1a3fee40-f7de-42df-b078-abde4117d47a-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.598138 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvctp\" (UniqueName: \"kubernetes.io/projected/4b7ebf6b-9957-4833-be7b-2123a7a456ee-kube-api-access-vvctp\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.600583 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a3fee40-f7de-42df-b078-abde4117d47a-kube-api-access-xljc6" (OuterVolumeSpecName: "kube-api-access-xljc6") pod "1a3fee40-f7de-42df-b078-abde4117d47a" (UID: "1a3fee40-f7de-42df-b078-abde4117d47a"). InnerVolumeSpecName "kube-api-access-xljc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.635913 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a3fee40-f7de-42df-b078-abde4117d47a-config-data" (OuterVolumeSpecName: "config-data") pod "1a3fee40-f7de-42df-b078-abde4117d47a" (UID: "1a3fee40-f7de-42df-b078-abde4117d47a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.640547 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a3fee40-f7de-42df-b078-abde4117d47a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1a3fee40-f7de-42df-b078-abde4117d47a" (UID: "1a3fee40-f7de-42df-b078-abde4117d47a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.700511 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r96p\" (UniqueName: \"kubernetes.io/projected/729e49b1-0bf4-43cf-b184-973dda23dc9c-kube-api-access-8r96p\") pod \"nova-cell1-conductor-0\" (UID: \"729e49b1-0bf4-43cf-b184-973dda23dc9c\") " pod="openstack/nova-cell1-conductor-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.700705 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729e49b1-0bf4-43cf-b184-973dda23dc9c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"729e49b1-0bf4-43cf-b184-973dda23dc9c\") " pod="openstack/nova-cell1-conductor-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.700748 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729e49b1-0bf4-43cf-b184-973dda23dc9c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"729e49b1-0bf4-43cf-b184-973dda23dc9c\") " pod="openstack/nova-cell1-conductor-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.700834 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a3fee40-f7de-42df-b078-abde4117d47a-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.700846 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xljc6\" (UniqueName: \"kubernetes.io/projected/1a3fee40-f7de-42df-b078-abde4117d47a-kube-api-access-xljc6\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.700858 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a3fee40-f7de-42df-b078-abde4117d47a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.706614 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/729e49b1-0bf4-43cf-b184-973dda23dc9c-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"729e49b1-0bf4-43cf-b184-973dda23dc9c\") " pod="openstack/nova-cell1-conductor-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.714502 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/729e49b1-0bf4-43cf-b184-973dda23dc9c-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"729e49b1-0bf4-43cf-b184-973dda23dc9c\") " pod="openstack/nova-cell1-conductor-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.716526 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r96p\" (UniqueName: \"kubernetes.io/projected/729e49b1-0bf4-43cf-b184-973dda23dc9c-kube-api-access-8r96p\") pod \"nova-cell1-conductor-0\" (UID: \"729e49b1-0bf4-43cf-b184-973dda23dc9c\") " pod="openstack/nova-cell1-conductor-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.718808 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:37:54 crc kubenswrapper[4930]: I0313 09:37:54.749666 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.097357 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"4b7ebf6b-9957-4833-be7b-2123a7a456ee","Type":"ContainerDied","Data":"c82dd72b332f72fdbb6883605d8f6273bfce4c53b9907e57f19ef3bc870e27c0"} Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.097755 4930 scope.go:117] "RemoveContainer" containerID="313f5fbaf4fd63adb12551298ae335950320ba02e956ae74e3d896436ab331d3" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.097389 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.100634 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.108708 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"1a3fee40-f7de-42df-b078-abde4117d47a","Type":"ContainerDied","Data":"030358db98d8e20d07411e9b9d1a3d94f161410d3a163098669a11735612c7bb"} Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.132641 4930 scope.go:117] "RemoveContainer" containerID="0e4ee155f43fa38e80c115a947e61020cf8bfca7893d21cf41942b070572ec31" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.242248 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.275682 4930 scope.go:117] "RemoveContainer" containerID="eb7012e87694f097a1abb477294439e571a1ae1c1cc305ae5a03627d60319a7b" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.277174 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.293552 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.312491 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.338390 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 09:37:55 crc kubenswrapper[4930]: E0313 09:37:55.339194 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a3fee40-f7de-42df-b078-abde4117d47a" containerName="nova-api-log" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.339286 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a3fee40-f7de-42df-b078-abde4117d47a" containerName="nova-api-log" Mar 13 09:37:55 crc kubenswrapper[4930]: E0313 09:37:55.339363 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a3fee40-f7de-42df-b078-abde4117d47a" containerName="nova-api-api" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.339457 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a3fee40-f7de-42df-b078-abde4117d47a" containerName="nova-api-api" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.339798 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a3fee40-f7de-42df-b078-abde4117d47a" containerName="nova-api-api" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.339912 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a3fee40-f7de-42df-b078-abde4117d47a" containerName="nova-api-log" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.349597 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.352608 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 09:37:55 crc kubenswrapper[4930]: W0313 09:37:55.362019 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod729e49b1_0bf4_43cf_b184_973dda23dc9c.slice/crio-385d2d67edfd3c1f0bed1f8b3fc39389e31fafc008b03c98e777fa01b54ad12a WatchSource:0}: Error finding container 385d2d67edfd3c1f0bed1f8b3fc39389e31fafc008b03c98e777fa01b54ad12a: Status 404 returned error can't find the container with id 385d2d67edfd3c1f0bed1f8b3fc39389e31fafc008b03c98e777fa01b54ad12a Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.373946 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.397897 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.411094 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.414713 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.416485 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.425785 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.436308 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aa2978c-5dda-4acf-8f73-ca0d8f987867-config-data\") pod \"nova-api-0\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.436658 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ndgr\" (UniqueName: \"kubernetes.io/projected/5aa2978c-5dda-4acf-8f73-ca0d8f987867-kube-api-access-5ndgr\") pod \"nova-api-0\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.436722 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aa2978c-5dda-4acf-8f73-ca0d8f987867-logs\") pod \"nova-api-0\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.436947 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa2978c-5dda-4acf-8f73-ca0d8f987867-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.438944 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.540278 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ndgr\" (UniqueName: \"kubernetes.io/projected/5aa2978c-5dda-4acf-8f73-ca0d8f987867-kube-api-access-5ndgr\") pod \"nova-api-0\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.540338 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aa2978c-5dda-4acf-8f73-ca0d8f987867-logs\") pod \"nova-api-0\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.540376 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5e802b-3388-4b21-9c96-7c3c925a35ea-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.540458 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa2978c-5dda-4acf-8f73-ca0d8f987867-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.540581 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5e802b-3388-4b21-9c96-7c3c925a35ea-config-data\") pod \"nova-scheduler-0\" (UID: \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.540862 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aa2978c-5dda-4acf-8f73-ca0d8f987867-config-data\") pod \"nova-api-0\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.540962 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rnhjp\" (UniqueName: \"kubernetes.io/projected/6b5e802b-3388-4b21-9c96-7c3c925a35ea-kube-api-access-rnhjp\") pod \"nova-scheduler-0\" (UID: \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.541002 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aa2978c-5dda-4acf-8f73-ca0d8f987867-logs\") pod \"nova-api-0\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.545368 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa2978c-5dda-4acf-8f73-ca0d8f987867-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.546316 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aa2978c-5dda-4acf-8f73-ca0d8f987867-config-data\") pod \"nova-api-0\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.560370 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ndgr\" (UniqueName: \"kubernetes.io/projected/5aa2978c-5dda-4acf-8f73-ca0d8f987867-kube-api-access-5ndgr\") pod \"nova-api-0\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.562222 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.642596 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rnhjp\" (UniqueName: \"kubernetes.io/projected/6b5e802b-3388-4b21-9c96-7c3c925a35ea-kube-api-access-rnhjp\") pod \"nova-scheduler-0\" (UID: \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.642691 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5e802b-3388-4b21-9c96-7c3c925a35ea-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.642752 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5e802b-3388-4b21-9c96-7c3c925a35ea-config-data\") pod \"nova-scheduler-0\" (UID: \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.647088 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5e802b-3388-4b21-9c96-7c3c925a35ea-config-data\") pod \"nova-scheduler-0\" (UID: \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.647190 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5e802b-3388-4b21-9c96-7c3c925a35ea-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.663002 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rnhjp\" (UniqueName: \"kubernetes.io/projected/6b5e802b-3388-4b21-9c96-7c3c925a35ea-kube-api-access-rnhjp\") pod \"nova-scheduler-0\" (UID: \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\") " pod="openstack/nova-scheduler-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.874225 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.986800 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a3fee40-f7de-42df-b078-abde4117d47a" path="/var/lib/kubelet/pods/1a3fee40-f7de-42df-b078-abde4117d47a/volumes" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.987790 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b7ebf6b-9957-4833-be7b-2123a7a456ee" path="/var/lib/kubelet/pods/4b7ebf6b-9957-4833-be7b-2123a7a456ee/volumes" Mar 13 09:37:55 crc kubenswrapper[4930]: I0313 09:37:55.990168 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68f691c8-ea38-424b-b373-1afbfdf9df4a" path="/var/lib/kubelet/pods/68f691c8-ea38-424b-b373-1afbfdf9df4a/volumes" Mar 13 09:37:56 crc kubenswrapper[4930]: I0313 09:37:56.080870 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:37:56 crc kubenswrapper[4930]: I0313 09:37:56.134413 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b45be8d9-5062-4d02-8a1b-3597f9c40c60","Type":"ContainerStarted","Data":"5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e"} Mar 13 09:37:56 crc kubenswrapper[4930]: I0313 09:37:56.134538 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b45be8d9-5062-4d02-8a1b-3597f9c40c60","Type":"ContainerStarted","Data":"52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875"} Mar 13 09:37:56 crc kubenswrapper[4930]: I0313 09:37:56.134551 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b45be8d9-5062-4d02-8a1b-3597f9c40c60","Type":"ContainerStarted","Data":"d5eef78a1cb0c7cfeb79659f2889956a2068971e640f1e303c5cff626ecc25fc"} Mar 13 09:37:56 crc kubenswrapper[4930]: I0313 09:37:56.166802 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.166782389 podStartE2EDuration="2.166782389s" podCreationTimestamp="2026-03-13 09:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:37:56.153159924 +0000 UTC m=+1516.903074621" watchObservedRunningTime="2026-03-13 09:37:56.166782389 +0000 UTC m=+1516.916697066" Mar 13 09:37:56 crc kubenswrapper[4930]: I0313 09:37:56.167416 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5aa2978c-5dda-4acf-8f73-ca0d8f987867","Type":"ContainerStarted","Data":"ccc1d4ad7e2eb04881f787d8e9d6e9673ec34da42c38b825c28a0327554d3781"} Mar 13 09:37:56 crc kubenswrapper[4930]: I0313 09:37:56.177760 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"729e49b1-0bf4-43cf-b184-973dda23dc9c","Type":"ContainerStarted","Data":"c1d7f1a04621ec8a2a124498df628471775c65808681a29614e5dd4b8e7207ea"} Mar 13 09:37:56 crc kubenswrapper[4930]: I0313 09:37:56.177805 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"729e49b1-0bf4-43cf-b184-973dda23dc9c","Type":"ContainerStarted","Data":"385d2d67edfd3c1f0bed1f8b3fc39389e31fafc008b03c98e777fa01b54ad12a"} Mar 13 09:37:56 crc kubenswrapper[4930]: I0313 09:37:56.179184 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 13 09:37:56 crc kubenswrapper[4930]: I0313 09:37:56.206550 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.206532399 podStartE2EDuration="2.206532399s" podCreationTimestamp="2026-03-13 09:37:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:37:56.193091197 +0000 UTC m=+1516.943005874" watchObservedRunningTime="2026-03-13 09:37:56.206532399 +0000 UTC m=+1516.956447076" Mar 13 09:37:56 crc kubenswrapper[4930]: I0313 09:37:56.596522 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:37:57 crc kubenswrapper[4930]: I0313 09:37:57.192003 4930 generic.go:334] "Generic (PLEG): container finished" podID="17fdbf0c-1335-4369-84b3-7db681759d19" containerID="08020393b80436228d130190c9583a49ddda5ae786b94588b37ff58a1ad73db6" exitCode=0 Mar 13 09:37:57 crc kubenswrapper[4930]: I0313 09:37:57.192118 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-dh2pc" event={"ID":"17fdbf0c-1335-4369-84b3-7db681759d19","Type":"ContainerDied","Data":"08020393b80436228d130190c9583a49ddda5ae786b94588b37ff58a1ad73db6"} Mar 13 09:37:57 crc kubenswrapper[4930]: I0313 09:37:57.195502 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5aa2978c-5dda-4acf-8f73-ca0d8f987867","Type":"ContainerStarted","Data":"7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74"} Mar 13 09:37:57 crc kubenswrapper[4930]: I0313 09:37:57.195558 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5aa2978c-5dda-4acf-8f73-ca0d8f987867","Type":"ContainerStarted","Data":"a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17"} Mar 13 09:37:57 crc kubenswrapper[4930]: I0313 09:37:57.197390 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b5e802b-3388-4b21-9c96-7c3c925a35ea","Type":"ContainerStarted","Data":"af306fc3c2a69f88b3beab2fbee7f37ddf999bc3988f0d277a1b07575d7d8bde"} Mar 13 09:37:57 crc kubenswrapper[4930]: I0313 09:37:57.197519 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b5e802b-3388-4b21-9c96-7c3c925a35ea","Type":"ContainerStarted","Data":"46a32e22bef2464bdb9a976cdf4918797b47040c6a939d0588bd678dffdd9cfd"} Mar 13 09:37:57 crc kubenswrapper[4930]: I0313 09:37:57.236899 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.236874157 podStartE2EDuration="2.236874157s" podCreationTimestamp="2026-03-13 09:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:37:57.235952973 +0000 UTC m=+1517.985867650" watchObservedRunningTime="2026-03-13 09:37:57.236874157 +0000 UTC m=+1517.986788834" Mar 13 09:37:57 crc kubenswrapper[4930]: I0313 09:37:57.272057 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.272027319 podStartE2EDuration="2.272027319s" podCreationTimestamp="2026-03-13 09:37:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:37:57.258301951 +0000 UTC m=+1518.008216628" watchObservedRunningTime="2026-03-13 09:37:57.272027319 +0000 UTC m=+1518.021941996" Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.700680 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.837823 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzvbw\" (UniqueName: \"kubernetes.io/projected/17fdbf0c-1335-4369-84b3-7db681759d19-kube-api-access-nzvbw\") pod \"17fdbf0c-1335-4369-84b3-7db681759d19\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.837876 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-config-data\") pod \"17fdbf0c-1335-4369-84b3-7db681759d19\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.838062 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-scripts\") pod \"17fdbf0c-1335-4369-84b3-7db681759d19\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.838240 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-combined-ca-bundle\") pod \"17fdbf0c-1335-4369-84b3-7db681759d19\" (UID: \"17fdbf0c-1335-4369-84b3-7db681759d19\") " Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.865829 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/17fdbf0c-1335-4369-84b3-7db681759d19-kube-api-access-nzvbw" (OuterVolumeSpecName: "kube-api-access-nzvbw") pod "17fdbf0c-1335-4369-84b3-7db681759d19" (UID: "17fdbf0c-1335-4369-84b3-7db681759d19"). InnerVolumeSpecName "kube-api-access-nzvbw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.866838 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-scripts" (OuterVolumeSpecName: "scripts") pod "17fdbf0c-1335-4369-84b3-7db681759d19" (UID: "17fdbf0c-1335-4369-84b3-7db681759d19"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.881563 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "17fdbf0c-1335-4369-84b3-7db681759d19" (UID: "17fdbf0c-1335-4369-84b3-7db681759d19"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.886647 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-config-data" (OuterVolumeSpecName: "config-data") pod "17fdbf0c-1335-4369-84b3-7db681759d19" (UID: "17fdbf0c-1335-4369-84b3-7db681759d19"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.940943 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzvbw\" (UniqueName: \"kubernetes.io/projected/17fdbf0c-1335-4369-84b3-7db681759d19-kube-api-access-nzvbw\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.940987 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.940997 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:58 crc kubenswrapper[4930]: I0313 09:37:58.941006 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/17fdbf0c-1335-4369-84b3-7db681759d19-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:37:59 crc kubenswrapper[4930]: I0313 09:37:59.222159 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-dh2pc" event={"ID":"17fdbf0c-1335-4369-84b3-7db681759d19","Type":"ContainerDied","Data":"84e981fbaa90967b8663bf2b9952c73c82fd4c03b24f07b79f3a25356e817ec7"} Mar 13 09:37:59 crc kubenswrapper[4930]: I0313 09:37:59.222226 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="84e981fbaa90967b8663bf2b9952c73c82fd4c03b24f07b79f3a25356e817ec7" Mar 13 09:37:59 crc kubenswrapper[4930]: I0313 09:37:59.222319 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-dh2pc" Mar 13 09:37:59 crc kubenswrapper[4930]: I0313 09:37:59.719752 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 09:37:59 crc kubenswrapper[4930]: I0313 09:37:59.720097 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.136455 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556578-qw9pg"] Mar 13 09:38:00 crc kubenswrapper[4930]: E0313 09:38:00.137014 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="17fdbf0c-1335-4369-84b3-7db681759d19" containerName="aodh-db-sync" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.137036 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="17fdbf0c-1335-4369-84b3-7db681759d19" containerName="aodh-db-sync" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.137286 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="17fdbf0c-1335-4369-84b3-7db681759d19" containerName="aodh-db-sync" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.138086 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556578-qw9pg" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.142869 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.143109 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.143132 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.154294 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556578-qw9pg"] Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.276409 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5dch\" (UniqueName: \"kubernetes.io/projected/a1645d8f-1162-44a6-a9df-9bb32862d73c-kube-api-access-m5dch\") pod \"auto-csr-approver-29556578-qw9pg\" (UID: \"a1645d8f-1162-44a6-a9df-9bb32862d73c\") " pod="openshift-infra/auto-csr-approver-29556578-qw9pg" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.318425 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-95hs2" podUID="0a86863c-63e1-4921-bd79-e6543a560043" containerName="registry-server" probeResult="failure" output=< Mar 13 09:38:00 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:38:00 crc kubenswrapper[4930]: > Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.379760 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5dch\" (UniqueName: \"kubernetes.io/projected/a1645d8f-1162-44a6-a9df-9bb32862d73c-kube-api-access-m5dch\") pod \"auto-csr-approver-29556578-qw9pg\" (UID: \"a1645d8f-1162-44a6-a9df-9bb32862d73c\") " pod="openshift-infra/auto-csr-approver-29556578-qw9pg" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.403970 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5dch\" (UniqueName: \"kubernetes.io/projected/a1645d8f-1162-44a6-a9df-9bb32862d73c-kube-api-access-m5dch\") pod \"auto-csr-approver-29556578-qw9pg\" (UID: \"a1645d8f-1162-44a6-a9df-9bb32862d73c\") " pod="openshift-infra/auto-csr-approver-29556578-qw9pg" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.418887 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.425713 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.429178 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.429766 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dc4s9" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.431873 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.442670 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.471552 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556578-qw9pg" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.595074 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-config-data\") pod \"aodh-0\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.595158 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-scripts\") pod \"aodh-0\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.595211 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-drhjt\" (UniqueName: \"kubernetes.io/projected/d856a1ba-f78b-4ec1-9d92-404464f8362c-kube-api-access-drhjt\") pod \"aodh-0\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.595843 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.700501 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.700630 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-config-data\") pod \"aodh-0\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.700662 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-scripts\") pod \"aodh-0\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.700690 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-drhjt\" (UniqueName: \"kubernetes.io/projected/d856a1ba-f78b-4ec1-9d92-404464f8362c-kube-api-access-drhjt\") pod \"aodh-0\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.711728 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-config-data\") pod \"aodh-0\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.712361 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-scripts\") pod \"aodh-0\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.715240 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.750034 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-drhjt\" (UniqueName: \"kubernetes.io/projected/d856a1ba-f78b-4ec1-9d92-404464f8362c-kube-api-access-drhjt\") pod \"aodh-0\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.801052 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 09:38:00 crc kubenswrapper[4930]: I0313 09:38:00.878936 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 09:38:01 crc kubenswrapper[4930]: W0313 09:38:01.246683 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda1645d8f_1162_44a6_a9df_9bb32862d73c.slice/crio-3415951ac17c34de8f4f26d494d8f3bad618661b23f01fb90b28e38d263da949 WatchSource:0}: Error finding container 3415951ac17c34de8f4f26d494d8f3bad618661b23f01fb90b28e38d263da949: Status 404 returned error can't find the container with id 3415951ac17c34de8f4f26d494d8f3bad618661b23f01fb90b28e38d263da949 Mar 13 09:38:01 crc kubenswrapper[4930]: I0313 09:38:01.253622 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556578-qw9pg"] Mar 13 09:38:01 crc kubenswrapper[4930]: I0313 09:38:01.444075 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 13 09:38:02 crc kubenswrapper[4930]: I0313 09:38:02.263805 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d856a1ba-f78b-4ec1-9d92-404464f8362c","Type":"ContainerStarted","Data":"1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6"} Mar 13 09:38:02 crc kubenswrapper[4930]: I0313 09:38:02.264150 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d856a1ba-f78b-4ec1-9d92-404464f8362c","Type":"ContainerStarted","Data":"f18329d9264e6864c435464e8f0d3e3caee6406ccd310d3699a5310ace8768d2"} Mar 13 09:38:02 crc kubenswrapper[4930]: I0313 09:38:02.277371 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556578-qw9pg" event={"ID":"a1645d8f-1162-44a6-a9df-9bb32862d73c","Type":"ContainerStarted","Data":"3415951ac17c34de8f4f26d494d8f3bad618661b23f01fb90b28e38d263da949"} Mar 13 09:38:03 crc kubenswrapper[4930]: I0313 09:38:03.307366 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556578-qw9pg" event={"ID":"a1645d8f-1162-44a6-a9df-9bb32862d73c","Type":"ContainerStarted","Data":"539a0f8a7b2944548c0bc24d36814e3b7804441ec50a4b08b916c45d1cde6c67"} Mar 13 09:38:03 crc kubenswrapper[4930]: I0313 09:38:03.326930 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556578-qw9pg" podStartSLOduration=2.484316391 podStartE2EDuration="3.326911004s" podCreationTimestamp="2026-03-13 09:38:00 +0000 UTC" firstStartedPulling="2026-03-13 09:38:01.249589543 +0000 UTC m=+1521.999504220" lastFinishedPulling="2026-03-13 09:38:02.092184166 +0000 UTC m=+1522.842098833" observedRunningTime="2026-03-13 09:38:03.324820021 +0000 UTC m=+1524.074734698" watchObservedRunningTime="2026-03-13 09:38:03.326911004 +0000 UTC m=+1524.076825681" Mar 13 09:38:03 crc kubenswrapper[4930]: I0313 09:38:03.991552 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:03 crc kubenswrapper[4930]: I0313 09:38:03.994986 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="sg-core" containerID="cri-o://716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df" gracePeriod=30 Mar 13 09:38:03 crc kubenswrapper[4930]: I0313 09:38:03.995156 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="proxy-httpd" containerID="cri-o://dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359" gracePeriod=30 Mar 13 09:38:03 crc kubenswrapper[4930]: I0313 09:38:03.995164 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="ceilometer-notification-agent" containerID="cri-o://aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77" gracePeriod=30 Mar 13 09:38:03 crc kubenswrapper[4930]: I0313 09:38:03.995253 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="ceilometer-central-agent" containerID="cri-o://93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36" gracePeriod=30 Mar 13 09:38:04 crc kubenswrapper[4930]: I0313 09:38:04.008281 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.1.2:3000/\": EOF" Mar 13 09:38:04 crc kubenswrapper[4930]: I0313 09:38:04.334413 4930 generic.go:334] "Generic (PLEG): container finished" podID="a1645d8f-1162-44a6-a9df-9bb32862d73c" containerID="539a0f8a7b2944548c0bc24d36814e3b7804441ec50a4b08b916c45d1cde6c67" exitCode=0 Mar 13 09:38:04 crc kubenswrapper[4930]: I0313 09:38:04.335478 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556578-qw9pg" event={"ID":"a1645d8f-1162-44a6-a9df-9bb32862d73c","Type":"ContainerDied","Data":"539a0f8a7b2944548c0bc24d36814e3b7804441ec50a4b08b916c45d1cde6c67"} Mar 13 09:38:04 crc kubenswrapper[4930]: I0313 09:38:04.341842 4930 generic.go:334] "Generic (PLEG): container finished" podID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerID="dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359" exitCode=0 Mar 13 09:38:04 crc kubenswrapper[4930]: I0313 09:38:04.341880 4930 generic.go:334] "Generic (PLEG): container finished" podID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerID="716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df" exitCode=2 Mar 13 09:38:04 crc kubenswrapper[4930]: I0313 09:38:04.341912 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce","Type":"ContainerDied","Data":"dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359"} Mar 13 09:38:04 crc kubenswrapper[4930]: I0313 09:38:04.341940 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce","Type":"ContainerDied","Data":"716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df"} Mar 13 09:38:04 crc kubenswrapper[4930]: I0313 09:38:04.412113 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Mar 13 09:38:04 crc kubenswrapper[4930]: I0313 09:38:04.720392 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 09:38:04 crc kubenswrapper[4930]: I0313 09:38:04.720923 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 09:38:04 crc kubenswrapper[4930]: I0313 09:38:04.796219 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.337948 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.364869 4930 generic.go:334] "Generic (PLEG): container finished" podID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerID="aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77" exitCode=0 Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.364902 4930 generic.go:334] "Generic (PLEG): container finished" podID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerID="93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36" exitCode=0 Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.364946 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce","Type":"ContainerDied","Data":"aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77"} Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.364978 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce","Type":"ContainerDied","Data":"93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36"} Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.364991 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce","Type":"ContainerDied","Data":"0a4543fcaac00419d9510241502cf601d76933d8c901ac2c0b62dd3787af495a"} Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.365010 4930 scope.go:117] "RemoveContainer" containerID="dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.365171 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.408782 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d856a1ba-f78b-4ec1-9d92-404464f8362c","Type":"ContainerStarted","Data":"9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734"} Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.456284 4930 scope.go:117] "RemoveContainer" containerID="716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.491259 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-log-httpd\") pod \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.491385 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-sg-core-conf-yaml\") pod \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.491512 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-scripts\") pod \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.491681 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-config-data\") pod \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.491781 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-njls8\" (UniqueName: \"kubernetes.io/projected/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-kube-api-access-njls8\") pod \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.491827 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-run-httpd\") pod \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.491878 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-combined-ca-bundle\") pod \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\" (UID: \"460dd2c5-ffb0-46d0-8f64-f07178a1c8ce\") " Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.492729 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" (UID: "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.493468 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" (UID: "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.512628 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-scripts" (OuterVolumeSpecName: "scripts") pod "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" (UID: "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.515830 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-kube-api-access-njls8" (OuterVolumeSpecName: "kube-api-access-njls8") pod "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" (UID: "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce"). InnerVolumeSpecName "kube-api-access-njls8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.524658 4930 scope.go:117] "RemoveContainer" containerID="aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.563993 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.564030 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.611015 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-njls8\" (UniqueName: \"kubernetes.io/projected/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-kube-api-access-njls8\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.611049 4930 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.611060 4930 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.611068 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.615872 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" (UID: "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.726325 4930 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.735744 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.5:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.736038 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.5:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.760787 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-config-data" (OuterVolumeSpecName: "config-data") pod "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" (UID: "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.856373 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.860895 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" (UID: "460dd2c5-ffb0-46d0-8f64-f07178a1c8ce"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.870124 4930 scope.go:117] "RemoveContainer" containerID="93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.878655 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.934629 4930 scope.go:117] "RemoveContainer" containerID="dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359" Mar 13 09:38:05 crc kubenswrapper[4930]: E0313 09:38:05.935083 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359\": container with ID starting with dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359 not found: ID does not exist" containerID="dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.935109 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359"} err="failed to get container status \"dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359\": rpc error: code = NotFound desc = could not find container \"dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359\": container with ID starting with dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359 not found: ID does not exist" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.935128 4930 scope.go:117] "RemoveContainer" containerID="716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df" Mar 13 09:38:05 crc kubenswrapper[4930]: E0313 09:38:05.935384 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df\": container with ID starting with 716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df not found: ID does not exist" containerID="716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.935410 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df"} err="failed to get container status \"716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df\": rpc error: code = NotFound desc = could not find container \"716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df\": container with ID starting with 716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df not found: ID does not exist" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.935425 4930 scope.go:117] "RemoveContainer" containerID="aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77" Mar 13 09:38:05 crc kubenswrapper[4930]: E0313 09:38:05.936019 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77\": container with ID starting with aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77 not found: ID does not exist" containerID="aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.936066 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77"} err="failed to get container status \"aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77\": rpc error: code = NotFound desc = could not find container \"aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77\": container with ID starting with aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77 not found: ID does not exist" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.936096 4930 scope.go:117] "RemoveContainer" containerID="93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36" Mar 13 09:38:05 crc kubenswrapper[4930]: E0313 09:38:05.939150 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36\": container with ID starting with 93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36 not found: ID does not exist" containerID="93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.939182 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36"} err="failed to get container status \"93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36\": rpc error: code = NotFound desc = could not find container \"93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36\": container with ID starting with 93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36 not found: ID does not exist" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.939198 4930 scope.go:117] "RemoveContainer" containerID="dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.940832 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359"} err="failed to get container status \"dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359\": rpc error: code = NotFound desc = could not find container \"dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359\": container with ID starting with dd65122a3f4547c5b4101a1b5cbbebcd64546e6744d917e58a8b860b7bcf3359 not found: ID does not exist" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.940856 4930 scope.go:117] "RemoveContainer" containerID="716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.941128 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df"} err="failed to get container status \"716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df\": rpc error: code = NotFound desc = could not find container \"716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df\": container with ID starting with 716a74c2ad80d2e6bf02a4e4de91c34910db783306657a81d6101c16e883a6df not found: ID does not exist" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.941144 4930 scope.go:117] "RemoveContainer" containerID="aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.941309 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77"} err="failed to get container status \"aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77\": rpc error: code = NotFound desc = could not find container \"aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77\": container with ID starting with aed8630fbe550e54d1479e539397062bcca47215bdb36cea747cf53a4f477e77 not found: ID does not exist" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.941322 4930 scope.go:117] "RemoveContainer" containerID="93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.941772 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36"} err="failed to get container status \"93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36\": rpc error: code = NotFound desc = could not find container \"93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36\": container with ID starting with 93d72e4cc070f8ce33abf00a2151af68e6ff4a0754ffc5a2997c559551992b36 not found: ID does not exist" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.958216 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.960228 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:05 crc kubenswrapper[4930]: I0313 09:38:05.970734 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556578-qw9pg" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.057516 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.063555 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m5dch\" (UniqueName: \"kubernetes.io/projected/a1645d8f-1162-44a6-a9df-9bb32862d73c-kube-api-access-m5dch\") pod \"a1645d8f-1162-44a6-a9df-9bb32862d73c\" (UID: \"a1645d8f-1162-44a6-a9df-9bb32862d73c\") " Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.085956 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.089355 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1645d8f-1162-44a6-a9df-9bb32862d73c-kube-api-access-m5dch" (OuterVolumeSpecName: "kube-api-access-m5dch") pod "a1645d8f-1162-44a6-a9df-9bb32862d73c" (UID: "a1645d8f-1162-44a6-a9df-9bb32862d73c"). InnerVolumeSpecName "kube-api-access-m5dch". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.099427 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:06 crc kubenswrapper[4930]: E0313 09:38:06.100244 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="proxy-httpd" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.100333 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="proxy-httpd" Mar 13 09:38:06 crc kubenswrapper[4930]: E0313 09:38:06.100391 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="sg-core" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.100459 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="sg-core" Mar 13 09:38:06 crc kubenswrapper[4930]: E0313 09:38:06.100571 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="ceilometer-notification-agent" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.100625 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="ceilometer-notification-agent" Mar 13 09:38:06 crc kubenswrapper[4930]: E0313 09:38:06.100685 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="ceilometer-central-agent" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.100734 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="ceilometer-central-agent" Mar 13 09:38:06 crc kubenswrapper[4930]: E0313 09:38:06.100801 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1645d8f-1162-44a6-a9df-9bb32862d73c" containerName="oc" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.100855 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1645d8f-1162-44a6-a9df-9bb32862d73c" containerName="oc" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.101110 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="ceilometer-notification-agent" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.101193 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1645d8f-1162-44a6-a9df-9bb32862d73c" containerName="oc" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.101278 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="proxy-httpd" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.101356 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="ceilometer-central-agent" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.101436 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" containerName="sg-core" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.105630 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.109132 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.109242 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.138085 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.166984 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lhs5v\" (UniqueName: \"kubernetes.io/projected/c117c905-135c-48f8-b8c8-3929b223faad-kube-api-access-lhs5v\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.167432 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.167613 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c117c905-135c-48f8-b8c8-3929b223faad-log-httpd\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.167764 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c117c905-135c-48f8-b8c8-3929b223faad-run-httpd\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.167836 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.167960 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-scripts\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.168321 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-config-data\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.168633 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m5dch\" (UniqueName: \"kubernetes.io/projected/a1645d8f-1162-44a6-a9df-9bb32862d73c-kube-api-access-m5dch\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.270553 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c117c905-135c-48f8-b8c8-3929b223faad-log-httpd\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.270637 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c117c905-135c-48f8-b8c8-3929b223faad-run-httpd\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.270663 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.270695 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-scripts\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.270778 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-config-data\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.270867 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lhs5v\" (UniqueName: \"kubernetes.io/projected/c117c905-135c-48f8-b8c8-3929b223faad-kube-api-access-lhs5v\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.270941 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.272343 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c117c905-135c-48f8-b8c8-3929b223faad-run-httpd\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.272496 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c117c905-135c-48f8-b8c8-3929b223faad-log-httpd\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.275605 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-config-data\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.275718 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.275999 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.285768 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-scripts\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.290088 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lhs5v\" (UniqueName: \"kubernetes.io/projected/c117c905-135c-48f8-b8c8-3929b223faad-kube-api-access-lhs5v\") pod \"ceilometer-0\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.408468 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556572-l9zlc"] Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.420494 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556572-l9zlc"] Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.426548 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.434660 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556578-qw9pg" event={"ID":"a1645d8f-1162-44a6-a9df-9bb32862d73c","Type":"ContainerDied","Data":"3415951ac17c34de8f4f26d494d8f3bad618661b23f01fb90b28e38d263da949"} Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.434706 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3415951ac17c34de8f4f26d494d8f3bad618661b23f01fb90b28e38d263da949" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.434776 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556578-qw9pg" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.498933 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.647001 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.1.7:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 09:38:06 crc kubenswrapper[4930]: I0313 09:38:06.647389 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.1.7:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 09:38:07 crc kubenswrapper[4930]: I0313 09:38:07.272405 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:07 crc kubenswrapper[4930]: I0313 09:38:07.606441 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:07 crc kubenswrapper[4930]: W0313 09:38:07.744716 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc117c905_135c_48f8_b8c8_3929b223faad.slice/crio-954d47f17dc1959b900c29fdfe55f852eb9964c8c71299145d6893dd5a361726 WatchSource:0}: Error finding container 954d47f17dc1959b900c29fdfe55f852eb9964c8c71299145d6893dd5a361726: Status 404 returned error can't find the container with id 954d47f17dc1959b900c29fdfe55f852eb9964c8c71299145d6893dd5a361726 Mar 13 09:38:08 crc kubenswrapper[4930]: I0313 09:38:08.000015 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27777941-a761-4e19-b4bf-6228b4e0221c" path="/var/lib/kubelet/pods/27777941-a761-4e19-b4bf-6228b4e0221c/volumes" Mar 13 09:38:08 crc kubenswrapper[4930]: I0313 09:38:08.007726 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="460dd2c5-ffb0-46d0-8f64-f07178a1c8ce" path="/var/lib/kubelet/pods/460dd2c5-ffb0-46d0-8f64-f07178a1c8ce/volumes" Mar 13 09:38:08 crc kubenswrapper[4930]: I0313 09:38:08.474452 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d856a1ba-f78b-4ec1-9d92-404464f8362c","Type":"ContainerStarted","Data":"ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47"} Mar 13 09:38:08 crc kubenswrapper[4930]: I0313 09:38:08.477525 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c117c905-135c-48f8-b8c8-3929b223faad","Type":"ContainerStarted","Data":"954d47f17dc1959b900c29fdfe55f852eb9964c8c71299145d6893dd5a361726"} Mar 13 09:38:09 crc kubenswrapper[4930]: I0313 09:38:09.314246 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:38:09 crc kubenswrapper[4930]: I0313 09:38:09.381524 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:38:09 crc kubenswrapper[4930]: I0313 09:38:09.489348 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c117c905-135c-48f8-b8c8-3929b223faad","Type":"ContainerStarted","Data":"e466bd0ecfa02c3c219d3b732074396d68943d23fb028f8bf449cdf6e7874c8a"} Mar 13 09:38:09 crc kubenswrapper[4930]: I0313 09:38:09.489453 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c117c905-135c-48f8-b8c8-3929b223faad","Type":"ContainerStarted","Data":"752d353d4364db5811751786323bc0f68568f85e367ea817db5b976d774856fe"} Mar 13 09:38:09 crc kubenswrapper[4930]: I0313 09:38:09.581365 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-95hs2"] Mar 13 09:38:10 crc kubenswrapper[4930]: I0313 09:38:10.502299 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-95hs2" podUID="0a86863c-63e1-4921-bd79-e6543a560043" containerName="registry-server" containerID="cri-o://6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be" gracePeriod=2 Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.262705 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.328830 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a86863c-63e1-4921-bd79-e6543a560043-catalog-content\") pod \"0a86863c-63e1-4921-bd79-e6543a560043\" (UID: \"0a86863c-63e1-4921-bd79-e6543a560043\") " Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.328982 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a86863c-63e1-4921-bd79-e6543a560043-utilities\") pod \"0a86863c-63e1-4921-bd79-e6543a560043\" (UID: \"0a86863c-63e1-4921-bd79-e6543a560043\") " Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.329261 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjkkv\" (UniqueName: \"kubernetes.io/projected/0a86863c-63e1-4921-bd79-e6543a560043-kube-api-access-xjkkv\") pod \"0a86863c-63e1-4921-bd79-e6543a560043\" (UID: \"0a86863c-63e1-4921-bd79-e6543a560043\") " Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.329657 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a86863c-63e1-4921-bd79-e6543a560043-utilities" (OuterVolumeSpecName: "utilities") pod "0a86863c-63e1-4921-bd79-e6543a560043" (UID: "0a86863c-63e1-4921-bd79-e6543a560043"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.330601 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0a86863c-63e1-4921-bd79-e6543a560043-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.334470 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a86863c-63e1-4921-bd79-e6543a560043-kube-api-access-xjkkv" (OuterVolumeSpecName: "kube-api-access-xjkkv") pod "0a86863c-63e1-4921-bd79-e6543a560043" (UID: "0a86863c-63e1-4921-bd79-e6543a560043"). InnerVolumeSpecName "kube-api-access-xjkkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.388762 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a86863c-63e1-4921-bd79-e6543a560043-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0a86863c-63e1-4921-bd79-e6543a560043" (UID: "0a86863c-63e1-4921-bd79-e6543a560043"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.432677 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjkkv\" (UniqueName: \"kubernetes.io/projected/0a86863c-63e1-4921-bd79-e6543a560043-kube-api-access-xjkkv\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.432709 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0a86863c-63e1-4921-bd79-e6543a560043-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.524164 4930 generic.go:334] "Generic (PLEG): container finished" podID="0a86863c-63e1-4921-bd79-e6543a560043" containerID="6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be" exitCode=0 Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.524232 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-95hs2" event={"ID":"0a86863c-63e1-4921-bd79-e6543a560043","Type":"ContainerDied","Data":"6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be"} Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.524260 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-95hs2" event={"ID":"0a86863c-63e1-4921-bd79-e6543a560043","Type":"ContainerDied","Data":"b24a76b4c42bac10ef90354044d285777921d357d0cb1a27ccedd02f8471bdc3"} Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.524279 4930 scope.go:117] "RemoveContainer" containerID="6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.524426 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-95hs2" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.535661 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d856a1ba-f78b-4ec1-9d92-404464f8362c","Type":"ContainerStarted","Data":"17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f"} Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.535844 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-api" containerID="cri-o://1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6" gracePeriod=30 Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.536317 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-listener" containerID="cri-o://17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f" gracePeriod=30 Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.536563 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-notifier" containerID="cri-o://ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47" gracePeriod=30 Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.536614 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-evaluator" containerID="cri-o://9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734" gracePeriod=30 Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.546677 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c117c905-135c-48f8-b8c8-3929b223faad","Type":"ContainerStarted","Data":"cf17ee87a5686b3c0a06100604431854492d1d4dd2621c1cb5ad5eee4a60fd4d"} Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.568100 4930 scope.go:117] "RemoveContainer" containerID="e8a12b74600495ac5afdbfa941792d415a79d4f129c347f079f29fe6d6d02978" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.587690 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.923728457 podStartE2EDuration="11.587669821s" podCreationTimestamp="2026-03-13 09:38:00 +0000 UTC" firstStartedPulling="2026-03-13 09:38:01.441385753 +0000 UTC m=+1522.191300430" lastFinishedPulling="2026-03-13 09:38:10.105327117 +0000 UTC m=+1530.855241794" observedRunningTime="2026-03-13 09:38:11.555944666 +0000 UTC m=+1532.305859343" watchObservedRunningTime="2026-03-13 09:38:11.587669821 +0000 UTC m=+1532.337584498" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.611481 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-95hs2"] Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.623951 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-95hs2"] Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.657679 4930 scope.go:117] "RemoveContainer" containerID="c42cbb1bbad0aee1ec5f69e1ac384be8a2a188a84624e825bf5f4748d1d47540" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.709095 4930 scope.go:117] "RemoveContainer" containerID="6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be" Mar 13 09:38:11 crc kubenswrapper[4930]: E0313 09:38:11.710010 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be\": container with ID starting with 6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be not found: ID does not exist" containerID="6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.710070 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be"} err="failed to get container status \"6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be\": rpc error: code = NotFound desc = could not find container \"6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be\": container with ID starting with 6c73c869e546202fe169209d64666dd9b3a0f7b1ad47a404a5cf0cf1a71a11be not found: ID does not exist" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.710103 4930 scope.go:117] "RemoveContainer" containerID="e8a12b74600495ac5afdbfa941792d415a79d4f129c347f079f29fe6d6d02978" Mar 13 09:38:11 crc kubenswrapper[4930]: E0313 09:38:11.712955 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8a12b74600495ac5afdbfa941792d415a79d4f129c347f079f29fe6d6d02978\": container with ID starting with e8a12b74600495ac5afdbfa941792d415a79d4f129c347f079f29fe6d6d02978 not found: ID does not exist" containerID="e8a12b74600495ac5afdbfa941792d415a79d4f129c347f079f29fe6d6d02978" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.712986 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a12b74600495ac5afdbfa941792d415a79d4f129c347f079f29fe6d6d02978"} err="failed to get container status \"e8a12b74600495ac5afdbfa941792d415a79d4f129c347f079f29fe6d6d02978\": rpc error: code = NotFound desc = could not find container \"e8a12b74600495ac5afdbfa941792d415a79d4f129c347f079f29fe6d6d02978\": container with ID starting with e8a12b74600495ac5afdbfa941792d415a79d4f129c347f079f29fe6d6d02978 not found: ID does not exist" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.713003 4930 scope.go:117] "RemoveContainer" containerID="c42cbb1bbad0aee1ec5f69e1ac384be8a2a188a84624e825bf5f4748d1d47540" Mar 13 09:38:11 crc kubenswrapper[4930]: E0313 09:38:11.714873 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c42cbb1bbad0aee1ec5f69e1ac384be8a2a188a84624e825bf5f4748d1d47540\": container with ID starting with c42cbb1bbad0aee1ec5f69e1ac384be8a2a188a84624e825bf5f4748d1d47540 not found: ID does not exist" containerID="c42cbb1bbad0aee1ec5f69e1ac384be8a2a188a84624e825bf5f4748d1d47540" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.714906 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c42cbb1bbad0aee1ec5f69e1ac384be8a2a188a84624e825bf5f4748d1d47540"} err="failed to get container status \"c42cbb1bbad0aee1ec5f69e1ac384be8a2a188a84624e825bf5f4748d1d47540\": rpc error: code = NotFound desc = could not find container \"c42cbb1bbad0aee1ec5f69e1ac384be8a2a188a84624e825bf5f4748d1d47540\": container with ID starting with c42cbb1bbad0aee1ec5f69e1ac384be8a2a188a84624e825bf5f4748d1d47540 not found: ID does not exist" Mar 13 09:38:11 crc kubenswrapper[4930]: I0313 09:38:11.989341 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a86863c-63e1-4921-bd79-e6543a560043" path="/var/lib/kubelet/pods/0a86863c-63e1-4921-bd79-e6543a560043/volumes" Mar 13 09:38:12 crc kubenswrapper[4930]: I0313 09:38:12.308792 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:38:12 crc kubenswrapper[4930]: I0313 09:38:12.309135 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:38:12 crc kubenswrapper[4930]: I0313 09:38:12.562106 4930 generic.go:334] "Generic (PLEG): container finished" podID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerID="ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47" exitCode=0 Mar 13 09:38:12 crc kubenswrapper[4930]: I0313 09:38:12.562135 4930 generic.go:334] "Generic (PLEG): container finished" podID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerID="9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734" exitCode=0 Mar 13 09:38:12 crc kubenswrapper[4930]: I0313 09:38:12.562143 4930 generic.go:334] "Generic (PLEG): container finished" podID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerID="1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6" exitCode=0 Mar 13 09:38:12 crc kubenswrapper[4930]: I0313 09:38:12.562160 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d856a1ba-f78b-4ec1-9d92-404464f8362c","Type":"ContainerDied","Data":"ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47"} Mar 13 09:38:12 crc kubenswrapper[4930]: I0313 09:38:12.562183 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d856a1ba-f78b-4ec1-9d92-404464f8362c","Type":"ContainerDied","Data":"9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734"} Mar 13 09:38:12 crc kubenswrapper[4930]: I0313 09:38:12.562191 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d856a1ba-f78b-4ec1-9d92-404464f8362c","Type":"ContainerDied","Data":"1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6"} Mar 13 09:38:13 crc kubenswrapper[4930]: I0313 09:38:13.575723 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c117c905-135c-48f8-b8c8-3929b223faad","Type":"ContainerStarted","Data":"d67799291b90df2ac475e83f58e18a7eb28e54ff4f579679de0d9e8952f8862d"} Mar 13 09:38:13 crc kubenswrapper[4930]: I0313 09:38:13.576075 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 09:38:13 crc kubenswrapper[4930]: I0313 09:38:13.575990 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="sg-core" containerID="cri-o://cf17ee87a5686b3c0a06100604431854492d1d4dd2621c1cb5ad5eee4a60fd4d" gracePeriod=30 Mar 13 09:38:13 crc kubenswrapper[4930]: I0313 09:38:13.575908 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="ceilometer-central-agent" containerID="cri-o://752d353d4364db5811751786323bc0f68568f85e367ea817db5b976d774856fe" gracePeriod=30 Mar 13 09:38:13 crc kubenswrapper[4930]: I0313 09:38:13.576065 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="ceilometer-notification-agent" containerID="cri-o://e466bd0ecfa02c3c219d3b732074396d68943d23fb028f8bf449cdf6e7874c8a" gracePeriod=30 Mar 13 09:38:13 crc kubenswrapper[4930]: I0313 09:38:13.576015 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="proxy-httpd" containerID="cri-o://d67799291b90df2ac475e83f58e18a7eb28e54ff4f579679de0d9e8952f8862d" gracePeriod=30 Mar 13 09:38:13 crc kubenswrapper[4930]: I0313 09:38:13.610825 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.7781727419999998 podStartE2EDuration="7.610810036s" podCreationTimestamp="2026-03-13 09:38:06 +0000 UTC" firstStartedPulling="2026-03-13 09:38:07.746827138 +0000 UTC m=+1528.496741815" lastFinishedPulling="2026-03-13 09:38:12.579464432 +0000 UTC m=+1533.329379109" observedRunningTime="2026-03-13 09:38:13.607475722 +0000 UTC m=+1534.357390399" watchObservedRunningTime="2026-03-13 09:38:13.610810036 +0000 UTC m=+1534.360724713" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.314633 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.419344 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-combined-ca-bundle\") pod \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\" (UID: \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\") " Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.419421 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-config-data\") pod \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\" (UID: \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\") " Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.419711 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7rf9\" (UniqueName: \"kubernetes.io/projected/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-kube-api-access-w7rf9\") pod \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\" (UID: \"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb\") " Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.426084 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-kube-api-access-w7rf9" (OuterVolumeSpecName: "kube-api-access-w7rf9") pod "cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb" (UID: "cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb"). InnerVolumeSpecName "kube-api-access-w7rf9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.452201 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-config-data" (OuterVolumeSpecName: "config-data") pod "cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb" (UID: "cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.457906 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb" (UID: "cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.522842 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7rf9\" (UniqueName: \"kubernetes.io/projected/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-kube-api-access-w7rf9\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.522880 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.522890 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.589786 4930 generic.go:334] "Generic (PLEG): container finished" podID="cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb" containerID="36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c" exitCode=137 Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.589836 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb","Type":"ContainerDied","Data":"36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c"} Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.589890 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb","Type":"ContainerDied","Data":"5c81fe72845fc14f5c6931fc16a77d95a67bfe25bd3d904c1ee6aac840aa3c93"} Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.589915 4930 scope.go:117] "RemoveContainer" containerID="36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.591662 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.593931 4930 generic.go:334] "Generic (PLEG): container finished" podID="c117c905-135c-48f8-b8c8-3929b223faad" containerID="d67799291b90df2ac475e83f58e18a7eb28e54ff4f579679de0d9e8952f8862d" exitCode=0 Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.593950 4930 generic.go:334] "Generic (PLEG): container finished" podID="c117c905-135c-48f8-b8c8-3929b223faad" containerID="cf17ee87a5686b3c0a06100604431854492d1d4dd2621c1cb5ad5eee4a60fd4d" exitCode=2 Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.593958 4930 generic.go:334] "Generic (PLEG): container finished" podID="c117c905-135c-48f8-b8c8-3929b223faad" containerID="e466bd0ecfa02c3c219d3b732074396d68943d23fb028f8bf449cdf6e7874c8a" exitCode=0 Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.593975 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c117c905-135c-48f8-b8c8-3929b223faad","Type":"ContainerDied","Data":"d67799291b90df2ac475e83f58e18a7eb28e54ff4f579679de0d9e8952f8862d"} Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.594001 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c117c905-135c-48f8-b8c8-3929b223faad","Type":"ContainerDied","Data":"cf17ee87a5686b3c0a06100604431854492d1d4dd2621c1cb5ad5eee4a60fd4d"} Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.594012 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c117c905-135c-48f8-b8c8-3929b223faad","Type":"ContainerDied","Data":"e466bd0ecfa02c3c219d3b732074396d68943d23fb028f8bf449cdf6e7874c8a"} Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.630252 4930 scope.go:117] "RemoveContainer" containerID="36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c" Mar 13 09:38:14 crc kubenswrapper[4930]: E0313 09:38:14.630911 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c\": container with ID starting with 36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c not found: ID does not exist" containerID="36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.630940 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c"} err="failed to get container status \"36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c\": rpc error: code = NotFound desc = could not find container \"36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c\": container with ID starting with 36e5dd27cdcae7561848788024a66967ae5843275ab9aa2c2d704ce1eff4ec3c not found: ID does not exist" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.635199 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.652624 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.665074 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 09:38:14 crc kubenswrapper[4930]: E0313 09:38:14.665909 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.665994 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 09:38:14 crc kubenswrapper[4930]: E0313 09:38:14.666083 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a86863c-63e1-4921-bd79-e6543a560043" containerName="extract-content" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.666140 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a86863c-63e1-4921-bd79-e6543a560043" containerName="extract-content" Mar 13 09:38:14 crc kubenswrapper[4930]: E0313 09:38:14.666229 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a86863c-63e1-4921-bd79-e6543a560043" containerName="registry-server" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.666300 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a86863c-63e1-4921-bd79-e6543a560043" containerName="registry-server" Mar 13 09:38:14 crc kubenswrapper[4930]: E0313 09:38:14.666386 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a86863c-63e1-4921-bd79-e6543a560043" containerName="extract-utilities" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.666544 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a86863c-63e1-4921-bd79-e6543a560043" containerName="extract-utilities" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.666847 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a86863c-63e1-4921-bd79-e6543a560043" containerName="registry-server" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.666953 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb" containerName="nova-cell1-novncproxy-novncproxy" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.671797 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.675607 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.684544 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.684715 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.684903 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.726454 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.727252 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.735641 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.739897 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a81b5d-7467-47b4-9f48-0c2d962c65bf-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.740004 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a81b5d-7467-47b4-9f48-0c2d962c65bf-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.740113 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a81b5d-7467-47b4-9f48-0c2d962c65bf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.740150 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a81b5d-7467-47b4-9f48-0c2d962c65bf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.740172 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc5wg\" (UniqueName: \"kubernetes.io/projected/10a81b5d-7467-47b4-9f48-0c2d962c65bf-kube-api-access-xc5wg\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.841744 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a81b5d-7467-47b4-9f48-0c2d962c65bf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.842023 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc5wg\" (UniqueName: \"kubernetes.io/projected/10a81b5d-7467-47b4-9f48-0c2d962c65bf-kube-api-access-xc5wg\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.842213 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a81b5d-7467-47b4-9f48-0c2d962c65bf-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.842827 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a81b5d-7467-47b4-9f48-0c2d962c65bf-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.842976 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a81b5d-7467-47b4-9f48-0c2d962c65bf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.847927 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a81b5d-7467-47b4-9f48-0c2d962c65bf-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.847948 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/10a81b5d-7467-47b4-9f48-0c2d962c65bf-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.848356 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10a81b5d-7467-47b4-9f48-0c2d962c65bf-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.851860 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10a81b5d-7467-47b4-9f48-0c2d962c65bf-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:14 crc kubenswrapper[4930]: I0313 09:38:14.862171 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc5wg\" (UniqueName: \"kubernetes.io/projected/10a81b5d-7467-47b4-9f48-0c2d962c65bf-kube-api-access-xc5wg\") pod \"nova-cell1-novncproxy-0\" (UID: \"10a81b5d-7467-47b4-9f48-0c2d962c65bf\") " pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.000226 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.517761 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 13 09:38:15 crc kubenswrapper[4930]: W0313 09:38:15.549911 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10a81b5d_7467_47b4_9f48_0c2d962c65bf.slice/crio-6b62d9c73d74a6cca63d5ef0cb9263400400bff3a330f30e644d55a234aaf8bf WatchSource:0}: Error finding container 6b62d9c73d74a6cca63d5ef0cb9263400400bff3a330f30e644d55a234aaf8bf: Status 404 returned error can't find the container with id 6b62d9c73d74a6cca63d5ef0cb9263400400bff3a330f30e644d55a234aaf8bf Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.570205 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.571028 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.571136 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.577706 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.624594 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"10a81b5d-7467-47b4-9f48-0c2d962c65bf","Type":"ContainerStarted","Data":"6b62d9c73d74a6cca63d5ef0cb9263400400bff3a330f30e644d55a234aaf8bf"} Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.637713 4930 generic.go:334] "Generic (PLEG): container finished" podID="c117c905-135c-48f8-b8c8-3929b223faad" containerID="752d353d4364db5811751786323bc0f68568f85e367ea817db5b976d774856fe" exitCode=0 Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.637910 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c117c905-135c-48f8-b8c8-3929b223faad","Type":"ContainerDied","Data":"752d353d4364db5811751786323bc0f68568f85e367ea817db5b976d774856fe"} Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.638958 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.643196 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.659992 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.867301 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-5cbxc"] Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.869920 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.894398 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-5cbxc"] Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.993371 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9sxr2\" (UniqueName: \"kubernetes.io/projected/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-kube-api-access-9sxr2\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.993884 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-config\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.994817 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb" path="/var/lib/kubelet/pods/cfb80ecb-f616-4aa9-9eb4-3e17e2f932cb/volumes" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.995878 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.996073 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.996260 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:15 crc kubenswrapper[4930]: I0313 09:38:15.996338 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.098168 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.098250 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.098416 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9sxr2\" (UniqueName: \"kubernetes.io/projected/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-kube-api-access-9sxr2\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.098525 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-config\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.099676 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.099914 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.100892 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-ovsdbserver-sb\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.101007 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-dns-swift-storage-0\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.101625 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-dns-svc\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.101893 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-ovsdbserver-nb\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.102612 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-config\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.123829 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9sxr2\" (UniqueName: \"kubernetes.io/projected/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-kube-api-access-9sxr2\") pod \"dnsmasq-dns-f84f9ccf-5cbxc\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.191698 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.275832 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.302859 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lhs5v\" (UniqueName: \"kubernetes.io/projected/c117c905-135c-48f8-b8c8-3929b223faad-kube-api-access-lhs5v\") pod \"c117c905-135c-48f8-b8c8-3929b223faad\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.302947 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-scripts\") pod \"c117c905-135c-48f8-b8c8-3929b223faad\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.305783 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-config-data\") pod \"c117c905-135c-48f8-b8c8-3929b223faad\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.305903 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-sg-core-conf-yaml\") pod \"c117c905-135c-48f8-b8c8-3929b223faad\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.306059 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c117c905-135c-48f8-b8c8-3929b223faad-run-httpd\") pod \"c117c905-135c-48f8-b8c8-3929b223faad\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.306347 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-combined-ca-bundle\") pod \"c117c905-135c-48f8-b8c8-3929b223faad\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.306386 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c117c905-135c-48f8-b8c8-3929b223faad-log-httpd\") pod \"c117c905-135c-48f8-b8c8-3929b223faad\" (UID: \"c117c905-135c-48f8-b8c8-3929b223faad\") " Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.306770 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c117c905-135c-48f8-b8c8-3929b223faad-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c117c905-135c-48f8-b8c8-3929b223faad" (UID: "c117c905-135c-48f8-b8c8-3929b223faad"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.307057 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c117c905-135c-48f8-b8c8-3929b223faad-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c117c905-135c-48f8-b8c8-3929b223faad" (UID: "c117c905-135c-48f8-b8c8-3929b223faad"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.307537 4930 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c117c905-135c-48f8-b8c8-3929b223faad-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.307554 4930 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c117c905-135c-48f8-b8c8-3929b223faad-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.310790 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c117c905-135c-48f8-b8c8-3929b223faad-kube-api-access-lhs5v" (OuterVolumeSpecName: "kube-api-access-lhs5v") pod "c117c905-135c-48f8-b8c8-3929b223faad" (UID: "c117c905-135c-48f8-b8c8-3929b223faad"). InnerVolumeSpecName "kube-api-access-lhs5v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.310801 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-scripts" (OuterVolumeSpecName: "scripts") pod "c117c905-135c-48f8-b8c8-3929b223faad" (UID: "c117c905-135c-48f8-b8c8-3929b223faad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.366519 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c117c905-135c-48f8-b8c8-3929b223faad" (UID: "c117c905-135c-48f8-b8c8-3929b223faad"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.409786 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lhs5v\" (UniqueName: \"kubernetes.io/projected/c117c905-135c-48f8-b8c8-3929b223faad-kube-api-access-lhs5v\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.409818 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.409829 4930 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.478604 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c117c905-135c-48f8-b8c8-3929b223faad" (UID: "c117c905-135c-48f8-b8c8-3929b223faad"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.513982 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.539011 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-config-data" (OuterVolumeSpecName: "config-data") pod "c117c905-135c-48f8-b8c8-3929b223faad" (UID: "c117c905-135c-48f8-b8c8-3929b223faad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.616166 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c117c905-135c-48f8-b8c8-3929b223faad-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.662252 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"10a81b5d-7467-47b4-9f48-0c2d962c65bf","Type":"ContainerStarted","Data":"3529f39e90d3524ed27f395daaee8a19ae7d70c38d171582c36db3a114e87b8f"} Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.669208 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.670222 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c117c905-135c-48f8-b8c8-3929b223faad","Type":"ContainerDied","Data":"954d47f17dc1959b900c29fdfe55f852eb9964c8c71299145d6893dd5a361726"} Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.670271 4930 scope.go:117] "RemoveContainer" containerID="d67799291b90df2ac475e83f58e18a7eb28e54ff4f579679de0d9e8952f8862d" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.695363 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.695344208 podStartE2EDuration="2.695344208s" podCreationTimestamp="2026-03-13 09:38:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:38:16.680917751 +0000 UTC m=+1537.430832428" watchObservedRunningTime="2026-03-13 09:38:16.695344208 +0000 UTC m=+1537.445258885" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.742270 4930 scope.go:117] "RemoveContainer" containerID="cf17ee87a5686b3c0a06100604431854492d1d4dd2621c1cb5ad5eee4a60fd4d" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.745557 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.765597 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.783538 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:16 crc kubenswrapper[4930]: E0313 09:38:16.784204 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="proxy-httpd" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.784228 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="proxy-httpd" Mar 13 09:38:16 crc kubenswrapper[4930]: E0313 09:38:16.784282 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="ceilometer-notification-agent" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.784294 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="ceilometer-notification-agent" Mar 13 09:38:16 crc kubenswrapper[4930]: E0313 09:38:16.784312 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="sg-core" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.784321 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="sg-core" Mar 13 09:38:16 crc kubenswrapper[4930]: E0313 09:38:16.784350 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="ceilometer-central-agent" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.784360 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="ceilometer-central-agent" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.784731 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="sg-core" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.784756 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="ceilometer-central-agent" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.784781 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="proxy-httpd" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.784798 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c117c905-135c-48f8-b8c8-3929b223faad" containerName="ceilometer-notification-agent" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.789105 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.798198 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.800034 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.804653 4930 scope.go:117] "RemoveContainer" containerID="e466bd0ecfa02c3c219d3b732074396d68943d23fb028f8bf449cdf6e7874c8a" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.816635 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.851409 4930 scope.go:117] "RemoveContainer" containerID="752d353d4364db5811751786323bc0f68568f85e367ea817db5b976d774856fe" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.888846 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-5cbxc"] Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.937352 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c061674b-0ea4-4acd-816d-b2946c459c4b-log-httpd\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.937899 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnjvq\" (UniqueName: \"kubernetes.io/projected/c061674b-0ea4-4acd-816d-b2946c459c4b-kube-api-access-pnjvq\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.938045 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-scripts\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.938073 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.938133 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c061674b-0ea4-4acd-816d-b2946c459c4b-run-httpd\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.938165 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-config-data\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:16 crc kubenswrapper[4930]: I0313 09:38:16.938247 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.040935 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-scripts\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.040979 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.041037 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c061674b-0ea4-4acd-816d-b2946c459c4b-run-httpd\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.041059 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-config-data\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.041077 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.041176 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c061674b-0ea4-4acd-816d-b2946c459c4b-log-httpd\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.041218 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnjvq\" (UniqueName: \"kubernetes.io/projected/c061674b-0ea4-4acd-816d-b2946c459c4b-kube-api-access-pnjvq\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.042837 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c061674b-0ea4-4acd-816d-b2946c459c4b-run-httpd\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.047406 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-scripts\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.048300 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c061674b-0ea4-4acd-816d-b2946c459c4b-log-httpd\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.050771 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-config-data\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.053357 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.085038 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.096556 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnjvq\" (UniqueName: \"kubernetes.io/projected/c061674b-0ea4-4acd-816d-b2946c459c4b-kube-api-access-pnjvq\") pod \"ceilometer-0\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.134352 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.680949 4930 generic.go:334] "Generic (PLEG): container finished" podID="c11761f9-55b2-4645-ae32-e2a4caaa8bb6" containerID="5d142d1645df31f75115da4a6211ab6b5bf9dacd0e7d33415b2c01edfbe54c2b" exitCode=0 Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.681248 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" event={"ID":"c11761f9-55b2-4645-ae32-e2a4caaa8bb6","Type":"ContainerDied","Data":"5d142d1645df31f75115da4a6211ab6b5bf9dacd0e7d33415b2c01edfbe54c2b"} Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.683404 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" event={"ID":"c11761f9-55b2-4645-ae32-e2a4caaa8bb6","Type":"ContainerStarted","Data":"5de233ac35bfb3ca3216fd1693bb31be51142d92255748e89ae9687bc6b36a6a"} Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.926266 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:17 crc kubenswrapper[4930]: I0313 09:38:17.993756 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c117c905-135c-48f8-b8c8-3929b223faad" path="/var/lib/kubelet/pods/c117c905-135c-48f8-b8c8-3929b223faad/volumes" Mar 13 09:38:18 crc kubenswrapper[4930]: I0313 09:38:18.696966 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c061674b-0ea4-4acd-816d-b2946c459c4b","Type":"ContainerStarted","Data":"41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92"} Mar 13 09:38:18 crc kubenswrapper[4930]: I0313 09:38:18.697563 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c061674b-0ea4-4acd-816d-b2946c459c4b","Type":"ContainerStarted","Data":"470796ac08e5562140b5d37056caee98c394cb1e565e35493eb0433c27846b2d"} Mar 13 09:38:18 crc kubenswrapper[4930]: I0313 09:38:18.699643 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" event={"ID":"c11761f9-55b2-4645-ae32-e2a4caaa8bb6","Type":"ContainerStarted","Data":"4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4"} Mar 13 09:38:18 crc kubenswrapper[4930]: I0313 09:38:18.699800 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:18 crc kubenswrapper[4930]: I0313 09:38:18.726351 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" podStartSLOduration=3.726324421 podStartE2EDuration="3.726324421s" podCreationTimestamp="2026-03-13 09:38:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:38:18.724693189 +0000 UTC m=+1539.474607876" watchObservedRunningTime="2026-03-13 09:38:18.726324421 +0000 UTC m=+1539.476239098" Mar 13 09:38:18 crc kubenswrapper[4930]: I0313 09:38:18.859534 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:38:18 crc kubenswrapper[4930]: I0313 09:38:18.859942 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" containerName="nova-api-log" containerID="cri-o://a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17" gracePeriod=30 Mar 13 09:38:18 crc kubenswrapper[4930]: I0313 09:38:18.860034 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" containerName="nova-api-api" containerID="cri-o://7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74" gracePeriod=30 Mar 13 09:38:19 crc kubenswrapper[4930]: I0313 09:38:19.521140 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:19 crc kubenswrapper[4930]: I0313 09:38:19.713553 4930 generic.go:334] "Generic (PLEG): container finished" podID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" containerID="a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17" exitCode=143 Mar 13 09:38:19 crc kubenswrapper[4930]: I0313 09:38:19.713626 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5aa2978c-5dda-4acf-8f73-ca0d8f987867","Type":"ContainerDied","Data":"a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17"} Mar 13 09:38:19 crc kubenswrapper[4930]: I0313 09:38:19.716502 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c061674b-0ea4-4acd-816d-b2946c459c4b","Type":"ContainerStarted","Data":"960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d"} Mar 13 09:38:20 crc kubenswrapper[4930]: I0313 09:38:20.000419 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:20 crc kubenswrapper[4930]: I0313 09:38:20.728410 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c061674b-0ea4-4acd-816d-b2946c459c4b","Type":"ContainerStarted","Data":"e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee"} Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.612210 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.692497 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aa2978c-5dda-4acf-8f73-ca0d8f987867-logs\") pod \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.692661 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa2978c-5dda-4acf-8f73-ca0d8f987867-combined-ca-bundle\") pod \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.692768 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aa2978c-5dda-4acf-8f73-ca0d8f987867-config-data\") pod \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.692848 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5ndgr\" (UniqueName: \"kubernetes.io/projected/5aa2978c-5dda-4acf-8f73-ca0d8f987867-kube-api-access-5ndgr\") pod \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\" (UID: \"5aa2978c-5dda-4acf-8f73-ca0d8f987867\") " Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.693075 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5aa2978c-5dda-4acf-8f73-ca0d8f987867-logs" (OuterVolumeSpecName: "logs") pod "5aa2978c-5dda-4acf-8f73-ca0d8f987867" (UID: "5aa2978c-5dda-4acf-8f73-ca0d8f987867"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.693701 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5aa2978c-5dda-4acf-8f73-ca0d8f987867-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.702741 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aa2978c-5dda-4acf-8f73-ca0d8f987867-kube-api-access-5ndgr" (OuterVolumeSpecName: "kube-api-access-5ndgr") pod "5aa2978c-5dda-4acf-8f73-ca0d8f987867" (UID: "5aa2978c-5dda-4acf-8f73-ca0d8f987867"). InnerVolumeSpecName "kube-api-access-5ndgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.742248 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aa2978c-5dda-4acf-8f73-ca0d8f987867-config-data" (OuterVolumeSpecName: "config-data") pod "5aa2978c-5dda-4acf-8f73-ca0d8f987867" (UID: "5aa2978c-5dda-4acf-8f73-ca0d8f987867"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.750388 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aa2978c-5dda-4acf-8f73-ca0d8f987867-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5aa2978c-5dda-4acf-8f73-ca0d8f987867" (UID: "5aa2978c-5dda-4acf-8f73-ca0d8f987867"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.755083 4930 generic.go:334] "Generic (PLEG): container finished" podID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" containerID="7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74" exitCode=0 Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.755166 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5aa2978c-5dda-4acf-8f73-ca0d8f987867","Type":"ContainerDied","Data":"7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74"} Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.755198 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"5aa2978c-5dda-4acf-8f73-ca0d8f987867","Type":"ContainerDied","Data":"ccc1d4ad7e2eb04881f787d8e9d6e9673ec34da42c38b825c28a0327554d3781"} Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.755220 4930 scope.go:117] "RemoveContainer" containerID="7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.755715 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.770394 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c061674b-0ea4-4acd-816d-b2946c459c4b","Type":"ContainerStarted","Data":"dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63"} Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.773600 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="ceilometer-central-agent" containerID="cri-o://41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92" gracePeriod=30 Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.774077 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.774141 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="proxy-httpd" containerID="cri-o://dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63" gracePeriod=30 Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.774187 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="sg-core" containerID="cri-o://e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee" gracePeriod=30 Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.774223 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="ceilometer-notification-agent" containerID="cri-o://960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d" gracePeriod=30 Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.797070 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aa2978c-5dda-4acf-8f73-ca0d8f987867-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.797102 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5aa2978c-5dda-4acf-8f73-ca0d8f987867-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.797111 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5ndgr\" (UniqueName: \"kubernetes.io/projected/5aa2978c-5dda-4acf-8f73-ca0d8f987867-kube-api-access-5ndgr\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.800275 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.8988146710000002 podStartE2EDuration="6.800263462s" podCreationTimestamp="2026-03-13 09:38:16 +0000 UTC" firstStartedPulling="2026-03-13 09:38:17.92393652 +0000 UTC m=+1538.673851197" lastFinishedPulling="2026-03-13 09:38:21.825385301 +0000 UTC m=+1542.575299988" observedRunningTime="2026-03-13 09:38:22.796755943 +0000 UTC m=+1543.546670620" watchObservedRunningTime="2026-03-13 09:38:22.800263462 +0000 UTC m=+1543.550178139" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.883221 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.913569 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.925790 4930 scope.go:117] "RemoveContainer" containerID="a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.931462 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 09:38:22 crc kubenswrapper[4930]: E0313 09:38:22.932003 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" containerName="nova-api-log" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.932021 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" containerName="nova-api-log" Mar 13 09:38:22 crc kubenswrapper[4930]: E0313 09:38:22.932101 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" containerName="nova-api-api" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.932109 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" containerName="nova-api-api" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.932321 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" containerName="nova-api-api" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.932357 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" containerName="nova-api-log" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.933730 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.937729 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.937859 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.938170 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 13 09:38:22 crc kubenswrapper[4930]: I0313 09:38:22.979750 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.001879 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-public-tls-certs\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.001992 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-config-data\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.002131 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f236668b-8609-40eb-bf77-3f64e6649be4-logs\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.002172 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.002188 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nwld\" (UniqueName: \"kubernetes.io/projected/f236668b-8609-40eb-bf77-3f64e6649be4-kube-api-access-4nwld\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.002263 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.047910 4930 scope.go:117] "RemoveContainer" containerID="7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74" Mar 13 09:38:23 crc kubenswrapper[4930]: E0313 09:38:23.048346 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74\": container with ID starting with 7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74 not found: ID does not exist" containerID="7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.048384 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74"} err="failed to get container status \"7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74\": rpc error: code = NotFound desc = could not find container \"7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74\": container with ID starting with 7ea61a298e348f99bbbd78756b74cb9b97251019a736d88a9a21a54e54034a74 not found: ID does not exist" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.048410 4930 scope.go:117] "RemoveContainer" containerID="a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17" Mar 13 09:38:23 crc kubenswrapper[4930]: E0313 09:38:23.048684 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17\": container with ID starting with a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17 not found: ID does not exist" containerID="a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.048711 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17"} err="failed to get container status \"a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17\": rpc error: code = NotFound desc = could not find container \"a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17\": container with ID starting with a97043634a4fc391b9fdf83db117dc647f77264e561b54656e62d48d54c8da17 not found: ID does not exist" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.104284 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f236668b-8609-40eb-bf77-3f64e6649be4-logs\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.104387 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.104411 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nwld\" (UniqueName: \"kubernetes.io/projected/f236668b-8609-40eb-bf77-3f64e6649be4-kube-api-access-4nwld\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.104593 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.104707 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-public-tls-certs\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.104945 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-config-data\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.107196 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f236668b-8609-40eb-bf77-3f64e6649be4-logs\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.122655 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-public-tls-certs\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.146230 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.149990 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-internal-tls-certs\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.150119 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nwld\" (UniqueName: \"kubernetes.io/projected/f236668b-8609-40eb-bf77-3f64e6649be4-kube-api-access-4nwld\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.153013 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-config-data\") pod \"nova-api-0\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.318514 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.799803 4930 generic.go:334] "Generic (PLEG): container finished" podID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerID="dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63" exitCode=0 Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.800131 4930 generic.go:334] "Generic (PLEG): container finished" podID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerID="e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee" exitCode=2 Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.800143 4930 generic.go:334] "Generic (PLEG): container finished" podID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerID="960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d" exitCode=0 Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.800189 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c061674b-0ea4-4acd-816d-b2946c459c4b","Type":"ContainerDied","Data":"dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63"} Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.800222 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c061674b-0ea4-4acd-816d-b2946c459c4b","Type":"ContainerDied","Data":"e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee"} Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.800234 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c061674b-0ea4-4acd-816d-b2946c459c4b","Type":"ContainerDied","Data":"960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d"} Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.856646 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.931664 4930 scope.go:117] "RemoveContainer" containerID="4abda406b7dbafb34ba62ab222a8d16f0082647d4e4d291a18bc2c69deca6f09" Mar 13 09:38:23 crc kubenswrapper[4930]: I0313 09:38:23.985531 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5aa2978c-5dda-4acf-8f73-ca0d8f987867" path="/var/lib/kubelet/pods/5aa2978c-5dda-4acf-8f73-ca0d8f987867/volumes" Mar 13 09:38:24 crc kubenswrapper[4930]: I0313 09:38:24.814955 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f236668b-8609-40eb-bf77-3f64e6649be4","Type":"ContainerStarted","Data":"b14214e032a6c54992a706af6ce3408ddbbe9dc2049cb8e8562f5046a2264b0a"} Mar 13 09:38:24 crc kubenswrapper[4930]: I0313 09:38:24.815574 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f236668b-8609-40eb-bf77-3f64e6649be4","Type":"ContainerStarted","Data":"c1af726f353830d35e4cdde637d9302be47587eb143bf54f1fd01563076b2e9f"} Mar 13 09:38:24 crc kubenswrapper[4930]: I0313 09:38:24.815591 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f236668b-8609-40eb-bf77-3f64e6649be4","Type":"ContainerStarted","Data":"87896f32d066d69547e40414b8022fba5965c0546c1e77aa546cdeaa5590266f"} Mar 13 09:38:24 crc kubenswrapper[4930]: I0313 09:38:24.837046 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.837026673 podStartE2EDuration="2.837026673s" podCreationTimestamp="2026-03-13 09:38:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:38:24.832401335 +0000 UTC m=+1545.582316012" watchObservedRunningTime="2026-03-13 09:38:24.837026673 +0000 UTC m=+1545.586941340" Mar 13 09:38:25 crc kubenswrapper[4930]: I0313 09:38:25.001213 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:25 crc kubenswrapper[4930]: I0313 09:38:25.023286 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:25 crc kubenswrapper[4930]: I0313 09:38:25.842651 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.024688 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-h78lz"] Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.026805 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.030425 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.030691 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.054796 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-h78lz"] Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.184406 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-scripts\") pod \"nova-cell1-cell-mapping-h78lz\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.184748 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-config-data\") pod \"nova-cell1-cell-mapping-h78lz\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.185085 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-h78lz\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.185243 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-47nq9\" (UniqueName: \"kubernetes.io/projected/b218ee1b-e381-4f43-b72b-8b11b440448f-kube-api-access-47nq9\") pod \"nova-cell1-cell-mapping-h78lz\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.278662 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.287932 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-scripts\") pod \"nova-cell1-cell-mapping-h78lz\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.287976 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-config-data\") pod \"nova-cell1-cell-mapping-h78lz\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.288135 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-h78lz\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.288192 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-47nq9\" (UniqueName: \"kubernetes.io/projected/b218ee1b-e381-4f43-b72b-8b11b440448f-kube-api-access-47nq9\") pod \"nova-cell1-cell-mapping-h78lz\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.296714 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-h78lz\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.296870 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-config-data\") pod \"nova-cell1-cell-mapping-h78lz\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.297964 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-scripts\") pod \"nova-cell1-cell-mapping-h78lz\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.322909 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-47nq9\" (UniqueName: \"kubernetes.io/projected/b218ee1b-e381-4f43-b72b-8b11b440448f-kube-api-access-47nq9\") pod \"nova-cell1-cell-mapping-h78lz\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.354070 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-t47jv"] Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.354333 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" podUID="ea17f053-2c9a-42ba-b729-106a9f9b90bb" containerName="dnsmasq-dns" containerID="cri-o://0b3811dabc8c56096d68904a6457198a1363ac79593afecedd9a2b99a45876ce" gracePeriod=10 Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.363258 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.898947 4930 generic.go:334] "Generic (PLEG): container finished" podID="ea17f053-2c9a-42ba-b729-106a9f9b90bb" containerID="0b3811dabc8c56096d68904a6457198a1363ac79593afecedd9a2b99a45876ce" exitCode=0 Mar 13 09:38:26 crc kubenswrapper[4930]: I0313 09:38:26.899475 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" event={"ID":"ea17f053-2c9a-42ba-b729-106a9f9b90bb","Type":"ContainerDied","Data":"0b3811dabc8c56096d68904a6457198a1363ac79593afecedd9a2b99a45876ce"} Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.225083 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-h78lz"] Mar 13 09:38:27 crc kubenswrapper[4930]: W0313 09:38:27.271312 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb218ee1b_e381_4f43_b72b_8b11b440448f.slice/crio-ffedbd5934bd8540cbc24b46a64dfad5eb82d6ab0ecf0ef9cb20b8de9168a9ba WatchSource:0}: Error finding container ffedbd5934bd8540cbc24b46a64dfad5eb82d6ab0ecf0ef9cb20b8de9168a9ba: Status 404 returned error can't find the container with id ffedbd5934bd8540cbc24b46a64dfad5eb82d6ab0ecf0ef9cb20b8de9168a9ba Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.277812 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.375230 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-config\") pod \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.375562 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-ovsdbserver-sb\") pod \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.375703 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-dns-swift-storage-0\") pod \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.375851 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-ovsdbserver-nb\") pod \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.375968 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bbwr\" (UniqueName: \"kubernetes.io/projected/ea17f053-2c9a-42ba-b729-106a9f9b90bb-kube-api-access-4bbwr\") pod \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.376094 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-dns-svc\") pod \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\" (UID: \"ea17f053-2c9a-42ba-b729-106a9f9b90bb\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.405218 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea17f053-2c9a-42ba-b729-106a9f9b90bb-kube-api-access-4bbwr" (OuterVolumeSpecName: "kube-api-access-4bbwr") pod "ea17f053-2c9a-42ba-b729-106a9f9b90bb" (UID: "ea17f053-2c9a-42ba-b729-106a9f9b90bb"). InnerVolumeSpecName "kube-api-access-4bbwr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.444090 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-config" (OuterVolumeSpecName: "config") pod "ea17f053-2c9a-42ba-b729-106a9f9b90bb" (UID: "ea17f053-2c9a-42ba-b729-106a9f9b90bb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.475167 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ea17f053-2c9a-42ba-b729-106a9f9b90bb" (UID: "ea17f053-2c9a-42ba-b729-106a9f9b90bb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.478178 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.478206 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.478220 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bbwr\" (UniqueName: \"kubernetes.io/projected/ea17f053-2c9a-42ba-b729-106a9f9b90bb-kube-api-access-4bbwr\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.494987 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ea17f053-2c9a-42ba-b729-106a9f9b90bb" (UID: "ea17f053-2c9a-42ba-b729-106a9f9b90bb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.508778 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ea17f053-2c9a-42ba-b729-106a9f9b90bb" (UID: "ea17f053-2c9a-42ba-b729-106a9f9b90bb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.550851 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ea17f053-2c9a-42ba-b729-106a9f9b90bb" (UID: "ea17f053-2c9a-42ba-b729-106a9f9b90bb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.580999 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.581345 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.581358 4930 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ea17f053-2c9a-42ba-b729-106a9f9b90bb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.831340 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.886663 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-combined-ca-bundle\") pod \"c061674b-0ea4-4acd-816d-b2946c459c4b\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.886726 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c061674b-0ea4-4acd-816d-b2946c459c4b-run-httpd\") pod \"c061674b-0ea4-4acd-816d-b2946c459c4b\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.886751 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-config-data\") pod \"c061674b-0ea4-4acd-816d-b2946c459c4b\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.886783 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c061674b-0ea4-4acd-816d-b2946c459c4b-log-httpd\") pod \"c061674b-0ea4-4acd-816d-b2946c459c4b\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.886861 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-sg-core-conf-yaml\") pod \"c061674b-0ea4-4acd-816d-b2946c459c4b\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.886927 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-scripts\") pod \"c061674b-0ea4-4acd-816d-b2946c459c4b\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.887009 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnjvq\" (UniqueName: \"kubernetes.io/projected/c061674b-0ea4-4acd-816d-b2946c459c4b-kube-api-access-pnjvq\") pod \"c061674b-0ea4-4acd-816d-b2946c459c4b\" (UID: \"c061674b-0ea4-4acd-816d-b2946c459c4b\") " Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.887531 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c061674b-0ea4-4acd-816d-b2946c459c4b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "c061674b-0ea4-4acd-816d-b2946c459c4b" (UID: "c061674b-0ea4-4acd-816d-b2946c459c4b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.888247 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c061674b-0ea4-4acd-816d-b2946c459c4b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "c061674b-0ea4-4acd-816d-b2946c459c4b" (UID: "c061674b-0ea4-4acd-816d-b2946c459c4b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.900324 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-scripts" (OuterVolumeSpecName: "scripts") pod "c061674b-0ea4-4acd-816d-b2946c459c4b" (UID: "c061674b-0ea4-4acd-816d-b2946c459c4b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.903978 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c061674b-0ea4-4acd-816d-b2946c459c4b-kube-api-access-pnjvq" (OuterVolumeSpecName: "kube-api-access-pnjvq") pod "c061674b-0ea4-4acd-816d-b2946c459c4b" (UID: "c061674b-0ea4-4acd-816d-b2946c459c4b"). InnerVolumeSpecName "kube-api-access-pnjvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.922797 4930 generic.go:334] "Generic (PLEG): container finished" podID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerID="41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92" exitCode=0 Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.923070 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c061674b-0ea4-4acd-816d-b2946c459c4b","Type":"ContainerDied","Data":"41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92"} Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.923352 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c061674b-0ea4-4acd-816d-b2946c459c4b","Type":"ContainerDied","Data":"470796ac08e5562140b5d37056caee98c394cb1e565e35493eb0433c27846b2d"} Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.923450 4930 scope.go:117] "RemoveContainer" containerID="dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.923143 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.929403 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" event={"ID":"ea17f053-2c9a-42ba-b729-106a9f9b90bb","Type":"ContainerDied","Data":"85c6a2bdea7e7230f6517d2c8c446febf96489e1f11f49ce94f527301039a663"} Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.929615 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-568d7fd7cf-t47jv" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.935061 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-h78lz" event={"ID":"b218ee1b-e381-4f43-b72b-8b11b440448f","Type":"ContainerStarted","Data":"c616c3f491277e469695830551caa9d391677e210c54ae9ecd436ba968511443"} Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.935108 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-h78lz" event={"ID":"b218ee1b-e381-4f43-b72b-8b11b440448f","Type":"ContainerStarted","Data":"ffedbd5934bd8540cbc24b46a64dfad5eb82d6ab0ecf0ef9cb20b8de9168a9ba"} Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.960211 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-h78lz" podStartSLOduration=2.960194585 podStartE2EDuration="2.960194585s" podCreationTimestamp="2026-03-13 09:38:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:38:27.95368021 +0000 UTC m=+1548.703594887" watchObservedRunningTime="2026-03-13 09:38:27.960194585 +0000 UTC m=+1548.710109262" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.976938 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "c061674b-0ea4-4acd-816d-b2946c459c4b" (UID: "c061674b-0ea4-4acd-816d-b2946c459c4b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.978924 4930 scope.go:117] "RemoveContainer" containerID="e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.991211 4930 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c061674b-0ea4-4acd-816d-b2946c459c4b-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.991883 4930 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c061674b-0ea4-4acd-816d-b2946c459c4b-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.991923 4930 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.991940 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:27 crc kubenswrapper[4930]: I0313 09:38:27.991955 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnjvq\" (UniqueName: \"kubernetes.io/projected/c061674b-0ea4-4acd-816d-b2946c459c4b-kube-api-access-pnjvq\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.012649 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-t47jv"] Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.022246 4930 scope.go:117] "RemoveContainer" containerID="960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.026951 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-568d7fd7cf-t47jv"] Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.055352 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c061674b-0ea4-4acd-816d-b2946c459c4b" (UID: "c061674b-0ea4-4acd-816d-b2946c459c4b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.079559 4930 scope.go:117] "RemoveContainer" containerID="41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.094806 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.097588 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-config-data" (OuterVolumeSpecName: "config-data") pod "c061674b-0ea4-4acd-816d-b2946c459c4b" (UID: "c061674b-0ea4-4acd-816d-b2946c459c4b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.129614 4930 scope.go:117] "RemoveContainer" containerID="dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63" Mar 13 09:38:28 crc kubenswrapper[4930]: E0313 09:38:28.130049 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63\": container with ID starting with dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63 not found: ID does not exist" containerID="dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.130094 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63"} err="failed to get container status \"dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63\": rpc error: code = NotFound desc = could not find container \"dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63\": container with ID starting with dd3294b5ce9c2e4fbb889fc374eed5110255ae12c5bf790d2afe1014fff1ba63 not found: ID does not exist" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.130119 4930 scope.go:117] "RemoveContainer" containerID="e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee" Mar 13 09:38:28 crc kubenswrapper[4930]: E0313 09:38:28.130320 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee\": container with ID starting with e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee not found: ID does not exist" containerID="e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.130368 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee"} err="failed to get container status \"e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee\": rpc error: code = NotFound desc = could not find container \"e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee\": container with ID starting with e80ee799e2beea5d3aa1d1d5f66645cf3469126e40146f2a9a0efaa3280accee not found: ID does not exist" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.130382 4930 scope.go:117] "RemoveContainer" containerID="960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d" Mar 13 09:38:28 crc kubenswrapper[4930]: E0313 09:38:28.130654 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d\": container with ID starting with 960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d not found: ID does not exist" containerID="960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.130680 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d"} err="failed to get container status \"960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d\": rpc error: code = NotFound desc = could not find container \"960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d\": container with ID starting with 960a0a4368a25f5f73ba292c18b1fdcc000aa495e4761dd631fcf685e96dd31d not found: ID does not exist" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.130698 4930 scope.go:117] "RemoveContainer" containerID="41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92" Mar 13 09:38:28 crc kubenswrapper[4930]: E0313 09:38:28.130907 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92\": container with ID starting with 41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92 not found: ID does not exist" containerID="41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.130929 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92"} err="failed to get container status \"41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92\": rpc error: code = NotFound desc = could not find container \"41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92\": container with ID starting with 41b18a74cd3fc44768fd0d4efd8ca7caa2358a335f6f2be2972357e4e2dc0f92 not found: ID does not exist" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.130946 4930 scope.go:117] "RemoveContainer" containerID="0b3811dabc8c56096d68904a6457198a1363ac79593afecedd9a2b99a45876ce" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.158651 4930 scope.go:117] "RemoveContainer" containerID="bcf4931f3f7a5bb25aaa59f905b399eb539d28b86e0927e1bc0c7721e688be5b" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.197866 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c061674b-0ea4-4acd-816d-b2946c459c4b-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.274818 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.288959 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.301424 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:28 crc kubenswrapper[4930]: E0313 09:38:28.302176 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="proxy-httpd" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.302201 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="proxy-httpd" Mar 13 09:38:28 crc kubenswrapper[4930]: E0313 09:38:28.302219 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="ceilometer-central-agent" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.302226 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="ceilometer-central-agent" Mar 13 09:38:28 crc kubenswrapper[4930]: E0313 09:38:28.302243 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="sg-core" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.302249 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="sg-core" Mar 13 09:38:28 crc kubenswrapper[4930]: E0313 09:38:28.302261 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea17f053-2c9a-42ba-b729-106a9f9b90bb" containerName="init" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.302268 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea17f053-2c9a-42ba-b729-106a9f9b90bb" containerName="init" Mar 13 09:38:28 crc kubenswrapper[4930]: E0313 09:38:28.302287 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea17f053-2c9a-42ba-b729-106a9f9b90bb" containerName="dnsmasq-dns" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.302293 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea17f053-2c9a-42ba-b729-106a9f9b90bb" containerName="dnsmasq-dns" Mar 13 09:38:28 crc kubenswrapper[4930]: E0313 09:38:28.302301 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="ceilometer-notification-agent" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.302308 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="ceilometer-notification-agent" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.302624 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="ceilometer-notification-agent" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.302645 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="proxy-httpd" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.302656 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="sg-core" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.302682 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" containerName="ceilometer-central-agent" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.302699 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea17f053-2c9a-42ba-b729-106a9f9b90bb" containerName="dnsmasq-dns" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.305203 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.307333 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.307974 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.327592 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.401899 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.402221 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-run-httpd\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.402525 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.402746 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-config-data\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.402893 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-log-httpd\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.403069 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-scripts\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.403245 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkg98\" (UniqueName: \"kubernetes.io/projected/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-kube-api-access-kkg98\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.505393 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.505457 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-config-data\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.505480 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-log-httpd\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.505541 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-scripts\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.505603 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkg98\" (UniqueName: \"kubernetes.io/projected/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-kube-api-access-kkg98\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.505695 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.505737 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-run-httpd\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.506263 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-run-httpd\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.506689 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-log-httpd\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.510162 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-scripts\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.511193 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.511498 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.516957 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-config-data\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.528295 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkg98\" (UniqueName: \"kubernetes.io/projected/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-kube-api-access-kkg98\") pod \"ceilometer-0\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " pod="openstack/ceilometer-0" Mar 13 09:38:28 crc kubenswrapper[4930]: I0313 09:38:28.713268 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:38:29 crc kubenswrapper[4930]: I0313 09:38:29.251952 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:38:29 crc kubenswrapper[4930]: I0313 09:38:29.985011 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c061674b-0ea4-4acd-816d-b2946c459c4b" path="/var/lib/kubelet/pods/c061674b-0ea4-4acd-816d-b2946c459c4b/volumes" Mar 13 09:38:29 crc kubenswrapper[4930]: I0313 09:38:29.986603 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea17f053-2c9a-42ba-b729-106a9f9b90bb" path="/var/lib/kubelet/pods/ea17f053-2c9a-42ba-b729-106a9f9b90bb/volumes" Mar 13 09:38:29 crc kubenswrapper[4930]: I0313 09:38:29.987374 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae1255e2-6e7d-43d0-ab12-2af33964e2f4","Type":"ContainerStarted","Data":"4165ef7f0faa1d759b6a7a390da07b24f88711c3df60cb5c5e9d1592f6ef6a16"} Mar 13 09:38:30 crc kubenswrapper[4930]: I0313 09:38:30.989710 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae1255e2-6e7d-43d0-ab12-2af33964e2f4","Type":"ContainerStarted","Data":"f05659e5fb73e35aa36f0b35a4df3bb3dcc03667b897b7e4af4fb48f282965b6"} Mar 13 09:38:30 crc kubenswrapper[4930]: I0313 09:38:30.990339 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae1255e2-6e7d-43d0-ab12-2af33964e2f4","Type":"ContainerStarted","Data":"58e9a1b30a4695ba81172383f9bc66a96743a5fcf919e2bb7a3a8df0b32b72a5"} Mar 13 09:38:32 crc kubenswrapper[4930]: I0313 09:38:32.016354 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae1255e2-6e7d-43d0-ab12-2af33964e2f4","Type":"ContainerStarted","Data":"e21f3ffcdceacd7569d61433c93fcf96498efdc5108290f59f34bdd7729e73b0"} Mar 13 09:38:33 crc kubenswrapper[4930]: I0313 09:38:33.319220 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 09:38:33 crc kubenswrapper[4930]: I0313 09:38:33.319843 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 09:38:34 crc kubenswrapper[4930]: I0313 09:38:34.052246 4930 generic.go:334] "Generic (PLEG): container finished" podID="b218ee1b-e381-4f43-b72b-8b11b440448f" containerID="c616c3f491277e469695830551caa9d391677e210c54ae9ecd436ba968511443" exitCode=0 Mar 13 09:38:34 crc kubenswrapper[4930]: I0313 09:38:34.053854 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-h78lz" event={"ID":"b218ee1b-e381-4f43-b72b-8b11b440448f","Type":"ContainerDied","Data":"c616c3f491277e469695830551caa9d391677e210c54ae9ecd436ba968511443"} Mar 13 09:38:34 crc kubenswrapper[4930]: I0313 09:38:34.057926 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae1255e2-6e7d-43d0-ab12-2af33964e2f4","Type":"ContainerStarted","Data":"2e485bbf244c2eee2edddbce5118e3f678745a96e30d7346f11f5ad5b42d573b"} Mar 13 09:38:34 crc kubenswrapper[4930]: I0313 09:38:34.058125 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 09:38:34 crc kubenswrapper[4930]: I0313 09:38:34.139706 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.449813772 podStartE2EDuration="6.139682252s" podCreationTimestamp="2026-03-13 09:38:28 +0000 UTC" firstStartedPulling="2026-03-13 09:38:29.256089705 +0000 UTC m=+1550.006004382" lastFinishedPulling="2026-03-13 09:38:32.945958185 +0000 UTC m=+1553.695872862" observedRunningTime="2026-03-13 09:38:34.111064386 +0000 UTC m=+1554.860979063" watchObservedRunningTime="2026-03-13 09:38:34.139682252 +0000 UTC m=+1554.889596929" Mar 13 09:38:34 crc kubenswrapper[4930]: I0313 09:38:34.342671 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f236668b-8609-40eb-bf77-3f64e6649be4" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.15:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 09:38:34 crc kubenswrapper[4930]: I0313 09:38:34.342686 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="f236668b-8609-40eb-bf77-3f64e6649be4" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.15:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.637224 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.783212 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-config-data\") pod \"b218ee1b-e381-4f43-b72b-8b11b440448f\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.783504 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-scripts\") pod \"b218ee1b-e381-4f43-b72b-8b11b440448f\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.783715 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-47nq9\" (UniqueName: \"kubernetes.io/projected/b218ee1b-e381-4f43-b72b-8b11b440448f-kube-api-access-47nq9\") pod \"b218ee1b-e381-4f43-b72b-8b11b440448f\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.783779 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-combined-ca-bundle\") pod \"b218ee1b-e381-4f43-b72b-8b11b440448f\" (UID: \"b218ee1b-e381-4f43-b72b-8b11b440448f\") " Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.789292 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-scripts" (OuterVolumeSpecName: "scripts") pod "b218ee1b-e381-4f43-b72b-8b11b440448f" (UID: "b218ee1b-e381-4f43-b72b-8b11b440448f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.790913 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b218ee1b-e381-4f43-b72b-8b11b440448f-kube-api-access-47nq9" (OuterVolumeSpecName: "kube-api-access-47nq9") pod "b218ee1b-e381-4f43-b72b-8b11b440448f" (UID: "b218ee1b-e381-4f43-b72b-8b11b440448f"). InnerVolumeSpecName "kube-api-access-47nq9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.815193 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b218ee1b-e381-4f43-b72b-8b11b440448f" (UID: "b218ee1b-e381-4f43-b72b-8b11b440448f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.827681 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-config-data" (OuterVolumeSpecName: "config-data") pod "b218ee1b-e381-4f43-b72b-8b11b440448f" (UID: "b218ee1b-e381-4f43-b72b-8b11b440448f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.887583 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-47nq9\" (UniqueName: \"kubernetes.io/projected/b218ee1b-e381-4f43-b72b-8b11b440448f-kube-api-access-47nq9\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.887646 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.887659 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:35 crc kubenswrapper[4930]: I0313 09:38:35.887691 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b218ee1b-e381-4f43-b72b-8b11b440448f-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:36 crc kubenswrapper[4930]: I0313 09:38:36.092788 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-h78lz" event={"ID":"b218ee1b-e381-4f43-b72b-8b11b440448f","Type":"ContainerDied","Data":"ffedbd5934bd8540cbc24b46a64dfad5eb82d6ab0ecf0ef9cb20b8de9168a9ba"} Mar 13 09:38:36 crc kubenswrapper[4930]: I0313 09:38:36.092828 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffedbd5934bd8540cbc24b46a64dfad5eb82d6ab0ecf0ef9cb20b8de9168a9ba" Mar 13 09:38:36 crc kubenswrapper[4930]: I0313 09:38:36.092853 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-h78lz" Mar 13 09:38:36 crc kubenswrapper[4930]: I0313 09:38:36.272771 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:38:36 crc kubenswrapper[4930]: I0313 09:38:36.273380 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f236668b-8609-40eb-bf77-3f64e6649be4" containerName="nova-api-log" containerID="cri-o://c1af726f353830d35e4cdde637d9302be47587eb143bf54f1fd01563076b2e9f" gracePeriod=30 Mar 13 09:38:36 crc kubenswrapper[4930]: I0313 09:38:36.273422 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="f236668b-8609-40eb-bf77-3f64e6649be4" containerName="nova-api-api" containerID="cri-o://b14214e032a6c54992a706af6ce3408ddbbe9dc2049cb8e8562f5046a2264b0a" gracePeriod=30 Mar 13 09:38:36 crc kubenswrapper[4930]: I0313 09:38:36.300955 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:38:36 crc kubenswrapper[4930]: I0313 09:38:36.302156 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6b5e802b-3388-4b21-9c96-7c3c925a35ea" containerName="nova-scheduler-scheduler" containerID="cri-o://af306fc3c2a69f88b3beab2fbee7f37ddf999bc3988f0d277a1b07575d7d8bde" gracePeriod=30 Mar 13 09:38:36 crc kubenswrapper[4930]: I0313 09:38:36.324798 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:38:36 crc kubenswrapper[4930]: I0313 09:38:36.325153 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerName="nova-metadata-log" containerID="cri-o://52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875" gracePeriod=30 Mar 13 09:38:36 crc kubenswrapper[4930]: I0313 09:38:36.325712 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerName="nova-metadata-metadata" containerID="cri-o://5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e" gracePeriod=30 Mar 13 09:38:37 crc kubenswrapper[4930]: I0313 09:38:37.105827 4930 generic.go:334] "Generic (PLEG): container finished" podID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerID="52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875" exitCode=143 Mar 13 09:38:37 crc kubenswrapper[4930]: I0313 09:38:37.105860 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b45be8d9-5062-4d02-8a1b-3597f9c40c60","Type":"ContainerDied","Data":"52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875"} Mar 13 09:38:37 crc kubenswrapper[4930]: I0313 09:38:37.109067 4930 generic.go:334] "Generic (PLEG): container finished" podID="f236668b-8609-40eb-bf77-3f64e6649be4" containerID="c1af726f353830d35e4cdde637d9302be47587eb143bf54f1fd01563076b2e9f" exitCode=143 Mar 13 09:38:37 crc kubenswrapper[4930]: I0313 09:38:37.109102 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f236668b-8609-40eb-bf77-3f64e6649be4","Type":"ContainerDied","Data":"c1af726f353830d35e4cdde637d9302be47587eb143bf54f1fd01563076b2e9f"} Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.043313 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.150479 4930 generic.go:334] "Generic (PLEG): container finished" podID="f236668b-8609-40eb-bf77-3f64e6649be4" containerID="b14214e032a6c54992a706af6ce3408ddbbe9dc2049cb8e8562f5046a2264b0a" exitCode=0 Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.150546 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f236668b-8609-40eb-bf77-3f64e6649be4","Type":"ContainerDied","Data":"b14214e032a6c54992a706af6ce3408ddbbe9dc2049cb8e8562f5046a2264b0a"} Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.152039 4930 generic.go:334] "Generic (PLEG): container finished" podID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerID="5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e" exitCode=0 Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.152064 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b45be8d9-5062-4d02-8a1b-3597f9c40c60","Type":"ContainerDied","Data":"5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e"} Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.152080 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"b45be8d9-5062-4d02-8a1b-3597f9c40c60","Type":"ContainerDied","Data":"d5eef78a1cb0c7cfeb79659f2889956a2068971e640f1e303c5cff626ecc25fc"} Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.152097 4930 scope.go:117] "RemoveContainer" containerID="5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.152223 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.194763 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-config-data\") pod \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.194821 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b45be8d9-5062-4d02-8a1b-3597f9c40c60-logs\") pod \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.195014 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-nova-metadata-tls-certs\") pod \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.195036 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-combined-ca-bundle\") pod \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.195092 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2crfk\" (UniqueName: \"kubernetes.io/projected/b45be8d9-5062-4d02-8a1b-3597f9c40c60-kube-api-access-2crfk\") pod \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\" (UID: \"b45be8d9-5062-4d02-8a1b-3597f9c40c60\") " Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.195511 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b45be8d9-5062-4d02-8a1b-3597f9c40c60-logs" (OuterVolumeSpecName: "logs") pod "b45be8d9-5062-4d02-8a1b-3597f9c40c60" (UID: "b45be8d9-5062-4d02-8a1b-3597f9c40c60"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.196118 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b45be8d9-5062-4d02-8a1b-3597f9c40c60-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.202480 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b45be8d9-5062-4d02-8a1b-3597f9c40c60-kube-api-access-2crfk" (OuterVolumeSpecName: "kube-api-access-2crfk") pod "b45be8d9-5062-4d02-8a1b-3597f9c40c60" (UID: "b45be8d9-5062-4d02-8a1b-3597f9c40c60"). InnerVolumeSpecName "kube-api-access-2crfk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.236807 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b45be8d9-5062-4d02-8a1b-3597f9c40c60" (UID: "b45be8d9-5062-4d02-8a1b-3597f9c40c60"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.239382 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.243250 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-config-data" (OuterVolumeSpecName: "config-data") pod "b45be8d9-5062-4d02-8a1b-3597f9c40c60" (UID: "b45be8d9-5062-4d02-8a1b-3597f9c40c60"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.252320 4930 scope.go:117] "RemoveContainer" containerID="52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.280230 4930 scope.go:117] "RemoveContainer" containerID="5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e" Mar 13 09:38:40 crc kubenswrapper[4930]: E0313 09:38:40.283779 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e\": container with ID starting with 5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e not found: ID does not exist" containerID="5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.283823 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e"} err="failed to get container status \"5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e\": rpc error: code = NotFound desc = could not find container \"5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e\": container with ID starting with 5efe0ad272d65e3df3feb45b0e41970a2a69f48be4be39cc9c517fbb5f1c5e4e not found: ID does not exist" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.283855 4930 scope.go:117] "RemoveContainer" containerID="52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875" Mar 13 09:38:40 crc kubenswrapper[4930]: E0313 09:38:40.284789 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875\": container with ID starting with 52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875 not found: ID does not exist" containerID="52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.284814 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875"} err="failed to get container status \"52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875\": rpc error: code = NotFound desc = could not find container \"52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875\": container with ID starting with 52467a146a2911b11be0c31c09e7da02e57423852b29d258a6f54fa39cd99875 not found: ID does not exist" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.299944 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "b45be8d9-5062-4d02-8a1b-3597f9c40c60" (UID: "b45be8d9-5062-4d02-8a1b-3597f9c40c60"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.300869 4930 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.300965 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.301068 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2crfk\" (UniqueName: \"kubernetes.io/projected/b45be8d9-5062-4d02-8a1b-3597f9c40c60-kube-api-access-2crfk\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.301154 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b45be8d9-5062-4d02-8a1b-3597f9c40c60-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.402111 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-combined-ca-bundle\") pod \"f236668b-8609-40eb-bf77-3f64e6649be4\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.402174 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-internal-tls-certs\") pod \"f236668b-8609-40eb-bf77-3f64e6649be4\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.402205 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4nwld\" (UniqueName: \"kubernetes.io/projected/f236668b-8609-40eb-bf77-3f64e6649be4-kube-api-access-4nwld\") pod \"f236668b-8609-40eb-bf77-3f64e6649be4\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.402242 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f236668b-8609-40eb-bf77-3f64e6649be4-logs\") pod \"f236668b-8609-40eb-bf77-3f64e6649be4\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.402395 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-config-data\") pod \"f236668b-8609-40eb-bf77-3f64e6649be4\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.402644 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-public-tls-certs\") pod \"f236668b-8609-40eb-bf77-3f64e6649be4\" (UID: \"f236668b-8609-40eb-bf77-3f64e6649be4\") " Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.403541 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f236668b-8609-40eb-bf77-3f64e6649be4-logs" (OuterVolumeSpecName: "logs") pod "f236668b-8609-40eb-bf77-3f64e6649be4" (UID: "f236668b-8609-40eb-bf77-3f64e6649be4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.405915 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f236668b-8609-40eb-bf77-3f64e6649be4-kube-api-access-4nwld" (OuterVolumeSpecName: "kube-api-access-4nwld") pod "f236668b-8609-40eb-bf77-3f64e6649be4" (UID: "f236668b-8609-40eb-bf77-3f64e6649be4"). InnerVolumeSpecName "kube-api-access-4nwld". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.433778 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-config-data" (OuterVolumeSpecName: "config-data") pod "f236668b-8609-40eb-bf77-3f64e6649be4" (UID: "f236668b-8609-40eb-bf77-3f64e6649be4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.435878 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f236668b-8609-40eb-bf77-3f64e6649be4" (UID: "f236668b-8609-40eb-bf77-3f64e6649be4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.463346 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "f236668b-8609-40eb-bf77-3f64e6649be4" (UID: "f236668b-8609-40eb-bf77-3f64e6649be4"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.465013 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f236668b-8609-40eb-bf77-3f64e6649be4" (UID: "f236668b-8609-40eb-bf77-3f64e6649be4"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.505367 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.505405 4930 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.505416 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.505425 4930 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f236668b-8609-40eb-bf77-3f64e6649be4-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.505453 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4nwld\" (UniqueName: \"kubernetes.io/projected/f236668b-8609-40eb-bf77-3f64e6649be4-kube-api-access-4nwld\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.505462 4930 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f236668b-8609-40eb-bf77-3f64e6649be4-logs\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.507615 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.526011 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.537606 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:38:40 crc kubenswrapper[4930]: E0313 09:38:40.538259 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b218ee1b-e381-4f43-b72b-8b11b440448f" containerName="nova-manage" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.538304 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b218ee1b-e381-4f43-b72b-8b11b440448f" containerName="nova-manage" Mar 13 09:38:40 crc kubenswrapper[4930]: E0313 09:38:40.538318 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f236668b-8609-40eb-bf77-3f64e6649be4" containerName="nova-api-api" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.538326 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f236668b-8609-40eb-bf77-3f64e6649be4" containerName="nova-api-api" Mar 13 09:38:40 crc kubenswrapper[4930]: E0313 09:38:40.538340 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerName="nova-metadata-metadata" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.538347 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerName="nova-metadata-metadata" Mar 13 09:38:40 crc kubenswrapper[4930]: E0313 09:38:40.538376 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerName="nova-metadata-log" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.538383 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerName="nova-metadata-log" Mar 13 09:38:40 crc kubenswrapper[4930]: E0313 09:38:40.538403 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f236668b-8609-40eb-bf77-3f64e6649be4" containerName="nova-api-log" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.538410 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f236668b-8609-40eb-bf77-3f64e6649be4" containerName="nova-api-log" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.538659 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f236668b-8609-40eb-bf77-3f64e6649be4" containerName="nova-api-log" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.538678 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f236668b-8609-40eb-bf77-3f64e6649be4" containerName="nova-api-api" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.538688 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerName="nova-metadata-metadata" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.538704 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b218ee1b-e381-4f43-b72b-8b11b440448f" containerName="nova-manage" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.538715 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerName="nova-metadata-log" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.553291 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.553402 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.557858 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.558158 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.712178 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca7da96e-e5ff-43fc-bca6-775f5aaba889-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.712224 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7da96e-e5ff-43fc-bca6-775f5aaba889-config-data\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.712359 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-stndv\" (UniqueName: \"kubernetes.io/projected/ca7da96e-e5ff-43fc-bca6-775f5aaba889-kube-api-access-stndv\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.712425 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7da96e-e5ff-43fc-bca6-775f5aaba889-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.712684 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca7da96e-e5ff-43fc-bca6-775f5aaba889-logs\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.816327 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca7da96e-e5ff-43fc-bca6-775f5aaba889-logs\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.816470 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca7da96e-e5ff-43fc-bca6-775f5aaba889-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.816499 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7da96e-e5ff-43fc-bca6-775f5aaba889-config-data\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.816573 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-stndv\" (UniqueName: \"kubernetes.io/projected/ca7da96e-e5ff-43fc-bca6-775f5aaba889-kube-api-access-stndv\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.816625 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7da96e-e5ff-43fc-bca6-775f5aaba889-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.817672 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ca7da96e-e5ff-43fc-bca6-775f5aaba889-logs\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.822203 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ca7da96e-e5ff-43fc-bca6-775f5aaba889-config-data\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.822711 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ca7da96e-e5ff-43fc-bca6-775f5aaba889-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.832250 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ca7da96e-e5ff-43fc-bca6-775f5aaba889-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.833647 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-stndv\" (UniqueName: \"kubernetes.io/projected/ca7da96e-e5ff-43fc-bca6-775f5aaba889-kube-api-access-stndv\") pod \"nova-metadata-0\" (UID: \"ca7da96e-e5ff-43fc-bca6-775f5aaba889\") " pod="openstack/nova-metadata-0" Mar 13 09:38:40 crc kubenswrapper[4930]: E0313 09:38:40.875422 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of af306fc3c2a69f88b3beab2fbee7f37ddf999bc3988f0d277a1b07575d7d8bde is running failed: container process not found" containerID="af306fc3c2a69f88b3beab2fbee7f37ddf999bc3988f0d277a1b07575d7d8bde" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 09:38:40 crc kubenswrapper[4930]: E0313 09:38:40.875760 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of af306fc3c2a69f88b3beab2fbee7f37ddf999bc3988f0d277a1b07575d7d8bde is running failed: container process not found" containerID="af306fc3c2a69f88b3beab2fbee7f37ddf999bc3988f0d277a1b07575d7d8bde" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 09:38:40 crc kubenswrapper[4930]: E0313 09:38:40.876076 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of af306fc3c2a69f88b3beab2fbee7f37ddf999bc3988f0d277a1b07575d7d8bde is running failed: container process not found" containerID="af306fc3c2a69f88b3beab2fbee7f37ddf999bc3988f0d277a1b07575d7d8bde" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 13 09:38:40 crc kubenswrapper[4930]: E0313 09:38:40.876147 4930 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of af306fc3c2a69f88b3beab2fbee7f37ddf999bc3988f0d277a1b07575d7d8bde is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="6b5e802b-3388-4b21-9c96-7c3c925a35ea" containerName="nova-scheduler-scheduler" Mar 13 09:38:40 crc kubenswrapper[4930]: I0313 09:38:40.934451 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.168025 4930 generic.go:334] "Generic (PLEG): container finished" podID="6b5e802b-3388-4b21-9c96-7c3c925a35ea" containerID="af306fc3c2a69f88b3beab2fbee7f37ddf999bc3988f0d277a1b07575d7d8bde" exitCode=0 Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.168071 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b5e802b-3388-4b21-9c96-7c3c925a35ea","Type":"ContainerDied","Data":"af306fc3c2a69f88b3beab2fbee7f37ddf999bc3988f0d277a1b07575d7d8bde"} Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.174176 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.174354 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"f236668b-8609-40eb-bf77-3f64e6649be4","Type":"ContainerDied","Data":"87896f32d066d69547e40414b8022fba5965c0546c1e77aa546cdeaa5590266f"} Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.174405 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.174417 4930 scope.go:117] "RemoveContainer" containerID="b14214e032a6c54992a706af6ce3408ddbbe9dc2049cb8e8562f5046a2264b0a" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.230214 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5e802b-3388-4b21-9c96-7c3c925a35ea-config-data\") pod \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\" (UID: \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\") " Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.230704 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnhjp\" (UniqueName: \"kubernetes.io/projected/6b5e802b-3388-4b21-9c96-7c3c925a35ea-kube-api-access-rnhjp\") pod \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\" (UID: \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\") " Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.230754 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5e802b-3388-4b21-9c96-7c3c925a35ea-combined-ca-bundle\") pod \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\" (UID: \"6b5e802b-3388-4b21-9c96-7c3c925a35ea\") " Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.253067 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b5e802b-3388-4b21-9c96-7c3c925a35ea-kube-api-access-rnhjp" (OuterVolumeSpecName: "kube-api-access-rnhjp") pod "6b5e802b-3388-4b21-9c96-7c3c925a35ea" (UID: "6b5e802b-3388-4b21-9c96-7c3c925a35ea"). InnerVolumeSpecName "kube-api-access-rnhjp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.254880 4930 scope.go:117] "RemoveContainer" containerID="c1af726f353830d35e4cdde637d9302be47587eb143bf54f1fd01563076b2e9f" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.278738 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5e802b-3388-4b21-9c96-7c3c925a35ea-config-data" (OuterVolumeSpecName: "config-data") pod "6b5e802b-3388-4b21-9c96-7c3c925a35ea" (UID: "6b5e802b-3388-4b21-9c96-7c3c925a35ea"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.283107 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.286674 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b5e802b-3388-4b21-9c96-7c3c925a35ea-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b5e802b-3388-4b21-9c96-7c3c925a35ea" (UID: "6b5e802b-3388-4b21-9c96-7c3c925a35ea"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.296174 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.320094 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 13 09:38:41 crc kubenswrapper[4930]: E0313 09:38:41.320660 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b5e802b-3388-4b21-9c96-7c3c925a35ea" containerName="nova-scheduler-scheduler" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.320681 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b5e802b-3388-4b21-9c96-7c3c925a35ea" containerName="nova-scheduler-scheduler" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.320908 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b5e802b-3388-4b21-9c96-7c3c925a35ea" containerName="nova-scheduler-scheduler" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.322292 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.331687 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.332234 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.332302 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.332017 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.332647 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-config-data\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.333018 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.333496 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-logs\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.333659 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.333783 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-public-tls-certs\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.333880 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdfbv\" (UniqueName: \"kubernetes.io/projected/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-kube-api-access-sdfbv\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.334183 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnhjp\" (UniqueName: \"kubernetes.io/projected/6b5e802b-3388-4b21-9c96-7c3c925a35ea-kube-api-access-rnhjp\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.334247 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b5e802b-3388-4b21-9c96-7c3c925a35ea-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.334308 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b5e802b-3388-4b21-9c96-7c3c925a35ea-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.435515 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-config-data\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.436213 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.437142 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-logs\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.437246 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.437359 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-public-tls-certs\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.437488 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdfbv\" (UniqueName: \"kubernetes.io/projected/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-kube-api-access-sdfbv\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.439268 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-logs\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.441554 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-internal-tls-certs\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.441717 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-config-data\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.441749 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-public-tls-certs\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.442562 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.453000 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdfbv\" (UniqueName: \"kubernetes.io/projected/94e0d64e-f0fc-4aff-a30f-c6033ba35fb3-kube-api-access-sdfbv\") pod \"nova-api-0\" (UID: \"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3\") " pod="openstack/nova-api-0" Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.565582 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 13 09:38:41 crc kubenswrapper[4930]: I0313 09:38:41.650451 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.023174 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" path="/var/lib/kubelet/pods/b45be8d9-5062-4d02-8a1b-3597f9c40c60/volumes" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.024539 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f236668b-8609-40eb-bf77-3f64e6649be4" path="/var/lib/kubelet/pods/f236668b-8609-40eb-bf77-3f64e6649be4/volumes" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.120675 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.212619 4930 generic.go:334] "Generic (PLEG): container finished" podID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerID="17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f" exitCode=137 Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.212688 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d856a1ba-f78b-4ec1-9d92-404464f8362c","Type":"ContainerDied","Data":"17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f"} Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.212828 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d856a1ba-f78b-4ec1-9d92-404464f8362c","Type":"ContainerDied","Data":"f18329d9264e6864c435464e8f0d3e3caee6406ccd310d3699a5310ace8768d2"} Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.212853 4930 scope.go:117] "RemoveContainer" containerID="17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.213025 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.231888 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6b5e802b-3388-4b21-9c96-7c3c925a35ea","Type":"ContainerDied","Data":"46a32e22bef2464bdb9a976cdf4918797b47040c6a939d0588bd678dffdd9cfd"} Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.232017 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.274126 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-drhjt\" (UniqueName: \"kubernetes.io/projected/d856a1ba-f78b-4ec1-9d92-404464f8362c-kube-api-access-drhjt\") pod \"d856a1ba-f78b-4ec1-9d92-404464f8362c\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.274300 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-scripts\") pod \"d856a1ba-f78b-4ec1-9d92-404464f8362c\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.274504 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-config-data\") pod \"d856a1ba-f78b-4ec1-9d92-404464f8362c\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.274625 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-combined-ca-bundle\") pod \"d856a1ba-f78b-4ec1-9d92-404464f8362c\" (UID: \"d856a1ba-f78b-4ec1-9d92-404464f8362c\") " Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.280611 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ca7da96e-e5ff-43fc-bca6-775f5aaba889","Type":"ContainerStarted","Data":"61df5eede218dff220942f1d440966a606072c416c212816d901c87ad5c74d71"} Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.280655 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ca7da96e-e5ff-43fc-bca6-775f5aaba889","Type":"ContainerStarted","Data":"902eb4788dd9e241f6e10fb6ae1d2c46d9d17a728bace7b6ffb06b1d0ba584c0"} Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.283168 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d856a1ba-f78b-4ec1-9d92-404464f8362c-kube-api-access-drhjt" (OuterVolumeSpecName: "kube-api-access-drhjt") pod "d856a1ba-f78b-4ec1-9d92-404464f8362c" (UID: "d856a1ba-f78b-4ec1-9d92-404464f8362c"). InnerVolumeSpecName "kube-api-access-drhjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.297627 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-scripts" (OuterVolumeSpecName: "scripts") pod "d856a1ba-f78b-4ec1-9d92-404464f8362c" (UID: "d856a1ba-f78b-4ec1-9d92-404464f8362c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.315587 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.315654 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.315698 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.316594 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.316643 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" gracePeriod=600 Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.338655 4930 scope.go:117] "RemoveContainer" containerID="ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.341544 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.377770 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-drhjt\" (UniqueName: \"kubernetes.io/projected/d856a1ba-f78b-4ec1-9d92-404464f8362c-kube-api-access-drhjt\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.378042 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.403485 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.454858 4930 scope.go:117] "RemoveContainer" containerID="9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.474884 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.513167 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:38:42 crc kubenswrapper[4930]: E0313 09:38:42.514313 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-api" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.514341 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-api" Mar 13 09:38:42 crc kubenswrapper[4930]: E0313 09:38:42.514393 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-evaluator" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.514408 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-evaluator" Mar 13 09:38:42 crc kubenswrapper[4930]: E0313 09:38:42.514420 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-notifier" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.514456 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-notifier" Mar 13 09:38:42 crc kubenswrapper[4930]: E0313 09:38:42.514495 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-listener" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.514504 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-listener" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.514970 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-evaluator" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.514991 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-api" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.515000 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-notifier" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.515048 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" containerName="aodh-listener" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.516387 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 09:38:42 crc kubenswrapper[4930]: E0313 09:38:42.525067 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.540257 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.547927 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.583785 4930 scope.go:117] "RemoveContainer" containerID="1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.595745 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68pvk\" (UniqueName: \"kubernetes.io/projected/b0c73ab8-6f27-4c43-8587-63e1d3fe0200-kube-api-access-68pvk\") pod \"nova-scheduler-0\" (UID: \"b0c73ab8-6f27-4c43-8587-63e1d3fe0200\") " pod="openstack/nova-scheduler-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.595812 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c73ab8-6f27-4c43-8587-63e1d3fe0200-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b0c73ab8-6f27-4c43-8587-63e1d3fe0200\") " pod="openstack/nova-scheduler-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.597696 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c73ab8-6f27-4c43-8587-63e1d3fe0200-config-data\") pod \"nova-scheduler-0\" (UID: \"b0c73ab8-6f27-4c43-8587-63e1d3fe0200\") " pod="openstack/nova-scheduler-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.699873 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c73ab8-6f27-4c43-8587-63e1d3fe0200-config-data\") pod \"nova-scheduler-0\" (UID: \"b0c73ab8-6f27-4c43-8587-63e1d3fe0200\") " pod="openstack/nova-scheduler-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.700236 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68pvk\" (UniqueName: \"kubernetes.io/projected/b0c73ab8-6f27-4c43-8587-63e1d3fe0200-kube-api-access-68pvk\") pod \"nova-scheduler-0\" (UID: \"b0c73ab8-6f27-4c43-8587-63e1d3fe0200\") " pod="openstack/nova-scheduler-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.700289 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c73ab8-6f27-4c43-8587-63e1d3fe0200-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b0c73ab8-6f27-4c43-8587-63e1d3fe0200\") " pod="openstack/nova-scheduler-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.704964 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b0c73ab8-6f27-4c43-8587-63e1d3fe0200-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"b0c73ab8-6f27-4c43-8587-63e1d3fe0200\") " pod="openstack/nova-scheduler-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.707489 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-config-data" (OuterVolumeSpecName: "config-data") pod "d856a1ba-f78b-4ec1-9d92-404464f8362c" (UID: "d856a1ba-f78b-4ec1-9d92-404464f8362c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.707886 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d856a1ba-f78b-4ec1-9d92-404464f8362c" (UID: "d856a1ba-f78b-4ec1-9d92-404464f8362c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.712138 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b0c73ab8-6f27-4c43-8587-63e1d3fe0200-config-data\") pod \"nova-scheduler-0\" (UID: \"b0c73ab8-6f27-4c43-8587-63e1d3fe0200\") " pod="openstack/nova-scheduler-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.717510 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68pvk\" (UniqueName: \"kubernetes.io/projected/b0c73ab8-6f27-4c43-8587-63e1d3fe0200-kube-api-access-68pvk\") pod \"nova-scheduler-0\" (UID: \"b0c73ab8-6f27-4c43-8587-63e1d3fe0200\") " pod="openstack/nova-scheduler-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.802993 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.803089 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d856a1ba-f78b-4ec1-9d92-404464f8362c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.819476 4930 scope.go:117] "RemoveContainer" containerID="17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f" Mar 13 09:38:42 crc kubenswrapper[4930]: E0313 09:38:42.819940 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f\": container with ID starting with 17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f not found: ID does not exist" containerID="17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.819993 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f"} err="failed to get container status \"17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f\": rpc error: code = NotFound desc = could not find container \"17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f\": container with ID starting with 17d43862f0655c4ae94d605f34ec46527973c87ee47a167dc7385d9db374c41f not found: ID does not exist" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.820033 4930 scope.go:117] "RemoveContainer" containerID="ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47" Mar 13 09:38:42 crc kubenswrapper[4930]: E0313 09:38:42.820340 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47\": container with ID starting with ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47 not found: ID does not exist" containerID="ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.820371 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47"} err="failed to get container status \"ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47\": rpc error: code = NotFound desc = could not find container \"ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47\": container with ID starting with ec3d2f8c5b5d68a88026240505fb8055d2c352f7efb85db46d0f664b31de6d47 not found: ID does not exist" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.820393 4930 scope.go:117] "RemoveContainer" containerID="9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734" Mar 13 09:38:42 crc kubenswrapper[4930]: E0313 09:38:42.822030 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734\": container with ID starting with 9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734 not found: ID does not exist" containerID="9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.822062 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734"} err="failed to get container status \"9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734\": rpc error: code = NotFound desc = could not find container \"9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734\": container with ID starting with 9cf9650295af8a41c37d12fc34a3c906bf4fdc3c21fe23e56ccb38b936f8f734 not found: ID does not exist" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.822084 4930 scope.go:117] "RemoveContainer" containerID="1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6" Mar 13 09:38:42 crc kubenswrapper[4930]: E0313 09:38:42.822861 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6\": container with ID starting with 1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6 not found: ID does not exist" containerID="1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.822889 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6"} err="failed to get container status \"1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6\": rpc error: code = NotFound desc = could not find container \"1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6\": container with ID starting with 1f03f98f9c0378fa4fcdd1b8947aa8e9df87a4edb733bc3de4cff584a8396dc6 not found: ID does not exist" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.822905 4930 scope.go:117] "RemoveContainer" containerID="af306fc3c2a69f88b3beab2fbee7f37ddf999bc3988f0d277a1b07575d7d8bde" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.826675 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.893740 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Mar 13 09:38:42 crc kubenswrapper[4930]: I0313 09:38:42.935946 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.036462 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.041247 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.074719 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.075119 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.075314 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.075493 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.078666 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.088233 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dc4s9" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.236417 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-public-tls-certs\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.236511 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jjswq\" (UniqueName: \"kubernetes.io/projected/d5419bb0-4585-43ee-9130-234ce8731cde-kube-api-access-jjswq\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.236645 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-config-data\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.236966 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-internal-tls-certs\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.237103 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-scripts\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.237287 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.302122 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3","Type":"ContainerStarted","Data":"16cd7bafa27863d9d2b8f6889c39c3ff60860d8b75b62ae3c12af1cb7f64f21e"} Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.302176 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3","Type":"ContainerStarted","Data":"cf6b10e37d29438dc8e27fb036f784dd423234f930d9b39c799bfeeb077d51ae"} Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.305717 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" exitCode=0 Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.305795 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766"} Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.305866 4930 scope.go:117] "RemoveContainer" containerID="9a600e0833fe6da30e3cb3376c93b675518bc4f00e36348b0e36bfb8a45c09d7" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.306725 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:38:43 crc kubenswrapper[4930]: E0313 09:38:43.307213 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.314766 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ca7da96e-e5ff-43fc-bca6-775f5aaba889","Type":"ContainerStarted","Data":"fd3df9cb98a754a3364f2cd59ab2b10cb19c08673a35cc83ad292c056be77abb"} Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.339412 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-internal-tls-certs\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.339522 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-scripts\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.339620 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.339649 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-public-tls-certs\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.339696 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jjswq\" (UniqueName: \"kubernetes.io/projected/d5419bb0-4585-43ee-9130-234ce8731cde-kube-api-access-jjswq\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.339817 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-config-data\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.344669 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-config-data\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.347776 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.348132 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-public-tls-certs\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.348999 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-internal-tls-certs\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.355590 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-scripts\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.366562 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jjswq\" (UniqueName: \"kubernetes.io/projected/d5419bb0-4585-43ee-9130-234ce8731cde-kube-api-access-jjswq\") pod \"aodh-0\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.379381 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.379363213 podStartE2EDuration="3.379363213s" podCreationTimestamp="2026-03-13 09:38:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:38:43.356610756 +0000 UTC m=+1564.106525423" watchObservedRunningTime="2026-03-13 09:38:43.379363213 +0000 UTC m=+1564.129277890" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.388003 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.424252 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 13 09:38:43 crc kubenswrapper[4930]: W0313 09:38:43.429582 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb0c73ab8_6f27_4c43_8587_63e1d3fe0200.slice/crio-20594a87e546ea446fc1e323b527b9fd3317aa9e80e13540adaa70fde996d09e WatchSource:0}: Error finding container 20594a87e546ea446fc1e323b527b9fd3317aa9e80e13540adaa70fde996d09e: Status 404 returned error can't find the container with id 20594a87e546ea446fc1e323b527b9fd3317aa9e80e13540adaa70fde996d09e Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.942931 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.984204 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b5e802b-3388-4b21-9c96-7c3c925a35ea" path="/var/lib/kubelet/pods/6b5e802b-3388-4b21-9c96-7c3c925a35ea/volumes" Mar 13 09:38:43 crc kubenswrapper[4930]: I0313 09:38:43.985093 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d856a1ba-f78b-4ec1-9d92-404464f8362c" path="/var/lib/kubelet/pods/d856a1ba-f78b-4ec1-9d92-404464f8362c/volumes" Mar 13 09:38:44 crc kubenswrapper[4930]: I0313 09:38:44.330001 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b0c73ab8-6f27-4c43-8587-63e1d3fe0200","Type":"ContainerStarted","Data":"c4a8a53ca67f1bae926ac3c0c836e87074e2191cb61d17da3c89ecf31f7deff2"} Mar 13 09:38:44 crc kubenswrapper[4930]: I0313 09:38:44.330056 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"b0c73ab8-6f27-4c43-8587-63e1d3fe0200","Type":"ContainerStarted","Data":"20594a87e546ea446fc1e323b527b9fd3317aa9e80e13540adaa70fde996d09e"} Mar 13 09:38:44 crc kubenswrapper[4930]: I0313 09:38:44.333983 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"94e0d64e-f0fc-4aff-a30f-c6033ba35fb3","Type":"ContainerStarted","Data":"be3bc98b3100c970f993a19401896af91d76eb59b9a29c5104e58be51fe7a812"} Mar 13 09:38:44 crc kubenswrapper[4930]: I0313 09:38:44.343664 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d5419bb0-4585-43ee-9130-234ce8731cde","Type":"ContainerStarted","Data":"cfbb96be6f9e279947c86fe9f71b554e91492cdeba927b9693f798fdcd663620"} Mar 13 09:38:44 crc kubenswrapper[4930]: I0313 09:38:44.349764 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.34974955 podStartE2EDuration="2.34974955s" podCreationTimestamp="2026-03-13 09:38:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:38:44.348056647 +0000 UTC m=+1565.097971314" watchObservedRunningTime="2026-03-13 09:38:44.34974955 +0000 UTC m=+1565.099664227" Mar 13 09:38:44 crc kubenswrapper[4930]: I0313 09:38:44.374558 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.374536219 podStartE2EDuration="3.374536219s" podCreationTimestamp="2026-03-13 09:38:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:38:44.367011138 +0000 UTC m=+1565.116925815" watchObservedRunningTime="2026-03-13 09:38:44.374536219 +0000 UTC m=+1565.124450896" Mar 13 09:38:44 crc kubenswrapper[4930]: I0313 09:38:44.721131 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.5:8775/\": dial tcp 10.217.1.5:8775: i/o timeout" Mar 13 09:38:44 crc kubenswrapper[4930]: I0313 09:38:44.721134 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="b45be8d9-5062-4d02-8a1b-3597f9c40c60" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.5:8775/\": dial tcp 10.217.1.5:8775: i/o timeout" Mar 13 09:38:45 crc kubenswrapper[4930]: I0313 09:38:45.355654 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d5419bb0-4585-43ee-9130-234ce8731cde","Type":"ContainerStarted","Data":"da5c96bf237888298b3a1b2948912a0b21740180ff441606d8c15b2a4abdb6d9"} Mar 13 09:38:45 crc kubenswrapper[4930]: I0313 09:38:45.935726 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 09:38:45 crc kubenswrapper[4930]: I0313 09:38:45.937015 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 13 09:38:46 crc kubenswrapper[4930]: I0313 09:38:46.367747 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d5419bb0-4585-43ee-9130-234ce8731cde","Type":"ContainerStarted","Data":"38c3ee5b46bba573a9741588b13094f9f6b96ca6958a7d7b5bb9367858bdfd17"} Mar 13 09:38:47 crc kubenswrapper[4930]: I0313 09:38:47.381736 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d5419bb0-4585-43ee-9130-234ce8731cde","Type":"ContainerStarted","Data":"ab3db3a80af457bd7b4a8dae8e2548f200e1ae5a50c6bdc6cecd064f95934dee"} Mar 13 09:38:47 crc kubenswrapper[4930]: I0313 09:38:47.382058 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d5419bb0-4585-43ee-9130-234ce8731cde","Type":"ContainerStarted","Data":"6750b989b595b6420d94d72e5e2309c5a1199f5e3192788cb02e421aef4949f6"} Mar 13 09:38:47 crc kubenswrapper[4930]: I0313 09:38:47.405281 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.472319242 podStartE2EDuration="5.405260035s" podCreationTimestamp="2026-03-13 09:38:42 +0000 UTC" firstStartedPulling="2026-03-13 09:38:43.947845676 +0000 UTC m=+1564.697760353" lastFinishedPulling="2026-03-13 09:38:46.880786469 +0000 UTC m=+1567.630701146" observedRunningTime="2026-03-13 09:38:47.402821313 +0000 UTC m=+1568.152735990" watchObservedRunningTime="2026-03-13 09:38:47.405260035 +0000 UTC m=+1568.155174712" Mar 13 09:38:47 crc kubenswrapper[4930]: I0313 09:38:47.828051 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 13 09:38:50 crc kubenswrapper[4930]: I0313 09:38:50.935191 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 09:38:50 crc kubenswrapper[4930]: I0313 09:38:50.935990 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 13 09:38:51 crc kubenswrapper[4930]: I0313 09:38:51.651309 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 09:38:51 crc kubenswrapper[4930]: I0313 09:38:51.651688 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 13 09:38:51 crc kubenswrapper[4930]: I0313 09:38:51.952654 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ca7da96e-e5ff-43fc-bca6-775f5aaba889" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.18:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 09:38:51 crc kubenswrapper[4930]: I0313 09:38:51.952687 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ca7da96e-e5ff-43fc-bca6-775f5aaba889" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.18:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 09:38:52 crc kubenswrapper[4930]: I0313 09:38:52.668633 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="94e0d64e-f0fc-4aff-a30f-c6033ba35fb3" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.1.19:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 09:38:52 crc kubenswrapper[4930]: I0313 09:38:52.668640 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="94e0d64e-f0fc-4aff-a30f-c6033ba35fb3" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.1.19:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 09:38:52 crc kubenswrapper[4930]: I0313 09:38:52.827976 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 13 09:38:52 crc kubenswrapper[4930]: I0313 09:38:52.864094 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 13 09:38:53 crc kubenswrapper[4930]: I0313 09:38:53.635201 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 13 09:38:57 crc kubenswrapper[4930]: I0313 09:38:57.971412 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:38:57 crc kubenswrapper[4930]: E0313 09:38:57.972202 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:38:58 crc kubenswrapper[4930]: I0313 09:38:58.719837 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 09:39:00 crc kubenswrapper[4930]: I0313 09:39:00.942954 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 09:39:00 crc kubenswrapper[4930]: I0313 09:39:00.944120 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 13 09:39:00 crc kubenswrapper[4930]: I0313 09:39:00.955977 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 09:39:00 crc kubenswrapper[4930]: I0313 09:39:00.957884 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 13 09:39:01 crc kubenswrapper[4930]: I0313 09:39:01.888612 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 09:39:01 crc kubenswrapper[4930]: I0313 09:39:01.889923 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 13 09:39:01 crc kubenswrapper[4930]: I0313 09:39:01.889977 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 09:39:01 crc kubenswrapper[4930]: I0313 09:39:01.902102 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 09:39:02 crc kubenswrapper[4930]: I0313 09:39:02.690919 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 13 09:39:02 crc kubenswrapper[4930]: I0313 09:39:02.698601 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 13 09:39:03 crc kubenswrapper[4930]: I0313 09:39:03.598128 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 09:39:03 crc kubenswrapper[4930]: I0313 09:39:03.598705 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="ecb88f7d-f623-4395-9c32-566a3ef1f0c9" containerName="kube-state-metrics" containerID="cri-o://30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f" gracePeriod=30 Mar 13 09:39:03 crc kubenswrapper[4930]: I0313 09:39:03.780605 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 13 09:39:03 crc kubenswrapper[4930]: I0313 09:39:03.780828 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="97bec685-87be-4c44-8678-189f13ffce7c" containerName="mysqld-exporter" containerID="cri-o://2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4" gracePeriod=30 Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.297966 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.385045 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v565g\" (UniqueName: \"kubernetes.io/projected/ecb88f7d-f623-4395-9c32-566a3ef1f0c9-kube-api-access-v565g\") pod \"ecb88f7d-f623-4395-9c32-566a3ef1f0c9\" (UID: \"ecb88f7d-f623-4395-9c32-566a3ef1f0c9\") " Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.407177 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecb88f7d-f623-4395-9c32-566a3ef1f0c9-kube-api-access-v565g" (OuterVolumeSpecName: "kube-api-access-v565g") pod "ecb88f7d-f623-4395-9c32-566a3ef1f0c9" (UID: "ecb88f7d-f623-4395-9c32-566a3ef1f0c9"). InnerVolumeSpecName "kube-api-access-v565g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.476862 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.487887 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v565g\" (UniqueName: \"kubernetes.io/projected/ecb88f7d-f623-4395-9c32-566a3ef1f0c9-kube-api-access-v565g\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.589660 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97bec685-87be-4c44-8678-189f13ffce7c-config-data\") pod \"97bec685-87be-4c44-8678-189f13ffce7c\" (UID: \"97bec685-87be-4c44-8678-189f13ffce7c\") " Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.589790 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bec685-87be-4c44-8678-189f13ffce7c-combined-ca-bundle\") pod \"97bec685-87be-4c44-8678-189f13ffce7c\" (UID: \"97bec685-87be-4c44-8678-189f13ffce7c\") " Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.590855 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzdb6\" (UniqueName: \"kubernetes.io/projected/97bec685-87be-4c44-8678-189f13ffce7c-kube-api-access-mzdb6\") pod \"97bec685-87be-4c44-8678-189f13ffce7c\" (UID: \"97bec685-87be-4c44-8678-189f13ffce7c\") " Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.598189 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97bec685-87be-4c44-8678-189f13ffce7c-kube-api-access-mzdb6" (OuterVolumeSpecName: "kube-api-access-mzdb6") pod "97bec685-87be-4c44-8678-189f13ffce7c" (UID: "97bec685-87be-4c44-8678-189f13ffce7c"). InnerVolumeSpecName "kube-api-access-mzdb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.643234 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97bec685-87be-4c44-8678-189f13ffce7c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "97bec685-87be-4c44-8678-189f13ffce7c" (UID: "97bec685-87be-4c44-8678-189f13ffce7c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.699659 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/97bec685-87be-4c44-8678-189f13ffce7c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.700020 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzdb6\" (UniqueName: \"kubernetes.io/projected/97bec685-87be-4c44-8678-189f13ffce7c-kube-api-access-mzdb6\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.709784 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97bec685-87be-4c44-8678-189f13ffce7c-config-data" (OuterVolumeSpecName: "config-data") pod "97bec685-87be-4c44-8678-189f13ffce7c" (UID: "97bec685-87be-4c44-8678-189f13ffce7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.743680 4930 generic.go:334] "Generic (PLEG): container finished" podID="97bec685-87be-4c44-8678-189f13ffce7c" containerID="2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4" exitCode=2 Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.743781 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"97bec685-87be-4c44-8678-189f13ffce7c","Type":"ContainerDied","Data":"2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4"} Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.743810 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"97bec685-87be-4c44-8678-189f13ffce7c","Type":"ContainerDied","Data":"083825c462c71eba24b362341b5110a57e69de30e1c7c2a5a481d14eaaaa1622"} Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.743826 4930 scope.go:117] "RemoveContainer" containerID="2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.743987 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.764500 4930 generic.go:334] "Generic (PLEG): container finished" podID="ecb88f7d-f623-4395-9c32-566a3ef1f0c9" containerID="30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f" exitCode=2 Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.764589 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.764609 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ecb88f7d-f623-4395-9c32-566a3ef1f0c9","Type":"ContainerDied","Data":"30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f"} Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.766553 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"ecb88f7d-f623-4395-9c32-566a3ef1f0c9","Type":"ContainerDied","Data":"f88eee0f524c5469e2db7689bf53be0693d5d333b8f7d542ad8c5a61cf5572b6"} Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.817327 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/97bec685-87be-4c44-8678-189f13ffce7c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.855461 4930 scope.go:117] "RemoveContainer" containerID="2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4" Mar 13 09:39:04 crc kubenswrapper[4930]: E0313 09:39:04.861136 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4\": container with ID starting with 2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4 not found: ID does not exist" containerID="2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.861190 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4"} err="failed to get container status \"2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4\": rpc error: code = NotFound desc = could not find container \"2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4\": container with ID starting with 2855a5cfcac7a52b049515a651bf289a30e8d7d89bf00b858b5cf278170e02e4 not found: ID does not exist" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.861217 4930 scope.go:117] "RemoveContainer" containerID="30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f" Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.896377 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.926004 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.961187 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 09:39:04 crc kubenswrapper[4930]: I0313 09:39:04.964643 4930 scope.go:117] "RemoveContainer" containerID="30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f" Mar 13 09:39:05 crc kubenswrapper[4930]: E0313 09:39:05.000056 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f\": container with ID starting with 30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f not found: ID does not exist" containerID="30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.000596 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f"} err="failed to get container status \"30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f\": rpc error: code = NotFound desc = could not find container \"30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f\": container with ID starting with 30af8dc50e4ac5360a87ee258b77fc0077ad1c22945271b55db64d0c5c15028f not found: ID does not exist" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.031788 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Mar 13 09:39:05 crc kubenswrapper[4930]: E0313 09:39:05.032551 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecb88f7d-f623-4395-9c32-566a3ef1f0c9" containerName="kube-state-metrics" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.032570 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecb88f7d-f623-4395-9c32-566a3ef1f0c9" containerName="kube-state-metrics" Mar 13 09:39:05 crc kubenswrapper[4930]: E0313 09:39:05.032628 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97bec685-87be-4c44-8678-189f13ffce7c" containerName="mysqld-exporter" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.032637 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="97bec685-87be-4c44-8678-189f13ffce7c" containerName="mysqld-exporter" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.033021 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecb88f7d-f623-4395-9c32-566a3ef1f0c9" containerName="kube-state-metrics" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.033045 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="97bec685-87be-4c44-8678-189f13ffce7c" containerName="mysqld-exporter" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.034398 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.050484 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.051531 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.052061 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.094903 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.106889 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.108597 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.110475 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.110565 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.126814 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.233272 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5\") " pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.233567 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79701ff1-8a11-4f24-9ab6-21f037545d5f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"79701ff1-8a11-4f24-9ab6-21f037545d5f\") " pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.233713 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj7fp\" (UniqueName: \"kubernetes.io/projected/cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5-kube-api-access-kj7fp\") pod \"mysqld-exporter-0\" (UID: \"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5\") " pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.234506 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5\") " pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.234728 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6hhdw\" (UniqueName: \"kubernetes.io/projected/79701ff1-8a11-4f24-9ab6-21f037545d5f-kube-api-access-6hhdw\") pod \"kube-state-metrics-0\" (UID: \"79701ff1-8a11-4f24-9ab6-21f037545d5f\") " pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.234845 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5-config-data\") pod \"mysqld-exporter-0\" (UID: \"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5\") " pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.234966 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/79701ff1-8a11-4f24-9ab6-21f037545d5f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"79701ff1-8a11-4f24-9ab6-21f037545d5f\") " pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.235131 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/79701ff1-8a11-4f24-9ab6-21f037545d5f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"79701ff1-8a11-4f24-9ab6-21f037545d5f\") " pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.338013 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5\") " pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.338113 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6hhdw\" (UniqueName: \"kubernetes.io/projected/79701ff1-8a11-4f24-9ab6-21f037545d5f-kube-api-access-6hhdw\") pod \"kube-state-metrics-0\" (UID: \"79701ff1-8a11-4f24-9ab6-21f037545d5f\") " pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.338171 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5-config-data\") pod \"mysqld-exporter-0\" (UID: \"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5\") " pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.338209 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/79701ff1-8a11-4f24-9ab6-21f037545d5f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"79701ff1-8a11-4f24-9ab6-21f037545d5f\") " pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.338277 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/79701ff1-8a11-4f24-9ab6-21f037545d5f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"79701ff1-8a11-4f24-9ab6-21f037545d5f\") " pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.338340 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5\") " pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.338426 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79701ff1-8a11-4f24-9ab6-21f037545d5f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"79701ff1-8a11-4f24-9ab6-21f037545d5f\") " pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.338537 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj7fp\" (UniqueName: \"kubernetes.io/projected/cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5-kube-api-access-kj7fp\") pod \"mysqld-exporter-0\" (UID: \"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5\") " pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.343409 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5\") " pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.343995 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5\") " pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.344196 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/79701ff1-8a11-4f24-9ab6-21f037545d5f-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"79701ff1-8a11-4f24-9ab6-21f037545d5f\") " pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.344614 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/79701ff1-8a11-4f24-9ab6-21f037545d5f-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"79701ff1-8a11-4f24-9ab6-21f037545d5f\") " pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.344901 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5-config-data\") pod \"mysqld-exporter-0\" (UID: \"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5\") " pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.346381 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/79701ff1-8a11-4f24-9ab6-21f037545d5f-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"79701ff1-8a11-4f24-9ab6-21f037545d5f\") " pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.364848 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj7fp\" (UniqueName: \"kubernetes.io/projected/cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5-kube-api-access-kj7fp\") pod \"mysqld-exporter-0\" (UID: \"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5\") " pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.366297 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6hhdw\" (UniqueName: \"kubernetes.io/projected/79701ff1-8a11-4f24-9ab6-21f037545d5f-kube-api-access-6hhdw\") pod \"kube-state-metrics-0\" (UID: \"79701ff1-8a11-4f24-9ab6-21f037545d5f\") " pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.381420 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.425335 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.953376 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.988751 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97bec685-87be-4c44-8678-189f13ffce7c" path="/var/lib/kubelet/pods/97bec685-87be-4c44-8678-189f13ffce7c/volumes" Mar 13 09:39:05 crc kubenswrapper[4930]: I0313 09:39:05.993183 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecb88f7d-f623-4395-9c32-566a3ef1f0c9" path="/var/lib/kubelet/pods/ecb88f7d-f623-4395-9c32-566a3ef1f0c9/volumes" Mar 13 09:39:06 crc kubenswrapper[4930]: W0313 09:39:06.042474 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod79701ff1_8a11_4f24_9ab6_21f037545d5f.slice/crio-1931925b7009c0471f1314689f25dfff355233a7120c7dc06150567a0ac64ee6 WatchSource:0}: Error finding container 1931925b7009c0471f1314689f25dfff355233a7120c7dc06150567a0ac64ee6: Status 404 returned error can't find the container with id 1931925b7009c0471f1314689f25dfff355233a7120c7dc06150567a0ac64ee6 Mar 13 09:39:06 crc kubenswrapper[4930]: I0313 09:39:06.049659 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 13 09:39:06 crc kubenswrapper[4930]: I0313 09:39:06.779863 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:39:06 crc kubenswrapper[4930]: I0313 09:39:06.780608 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="ceilometer-central-agent" containerID="cri-o://58e9a1b30a4695ba81172383f9bc66a96743a5fcf919e2bb7a3a8df0b32b72a5" gracePeriod=30 Mar 13 09:39:06 crc kubenswrapper[4930]: I0313 09:39:06.780772 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="proxy-httpd" containerID="cri-o://2e485bbf244c2eee2edddbce5118e3f678745a96e30d7346f11f5ad5b42d573b" gracePeriod=30 Mar 13 09:39:06 crc kubenswrapper[4930]: I0313 09:39:06.780831 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="sg-core" containerID="cri-o://e21f3ffcdceacd7569d61433c93fcf96498efdc5108290f59f34bdd7729e73b0" gracePeriod=30 Mar 13 09:39:06 crc kubenswrapper[4930]: I0313 09:39:06.781019 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="ceilometer-notification-agent" containerID="cri-o://f05659e5fb73e35aa36f0b35a4df3bb3dcc03667b897b7e4af4fb48f282965b6" gracePeriod=30 Mar 13 09:39:06 crc kubenswrapper[4930]: I0313 09:39:06.830852 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"79701ff1-8a11-4f24-9ab6-21f037545d5f","Type":"ContainerStarted","Data":"1931925b7009c0471f1314689f25dfff355233a7120c7dc06150567a0ac64ee6"} Mar 13 09:39:06 crc kubenswrapper[4930]: I0313 09:39:06.832471 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5","Type":"ContainerStarted","Data":"3ab844ac9eb56275b815474c53be2710cc7cbab2ea1cbed8b20d6903967b36e2"} Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.889918 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"cfbb3b1a-5b37-4ab2-be3e-bf9ef018c3a5","Type":"ContainerStarted","Data":"5358b2253f622887d593e3bc765eee13bca5fcdf0ae8adb4a1c7def8a82acbce"} Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.924859 4930 generic.go:334] "Generic (PLEG): container finished" podID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerID="2e485bbf244c2eee2edddbce5118e3f678745a96e30d7346f11f5ad5b42d573b" exitCode=0 Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.925180 4930 generic.go:334] "Generic (PLEG): container finished" podID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerID="e21f3ffcdceacd7569d61433c93fcf96498efdc5108290f59f34bdd7729e73b0" exitCode=2 Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.925260 4930 generic.go:334] "Generic (PLEG): container finished" podID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerID="f05659e5fb73e35aa36f0b35a4df3bb3dcc03667b897b7e4af4fb48f282965b6" exitCode=0 Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.925346 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae1255e2-6e7d-43d0-ab12-2af33964e2f4","Type":"ContainerDied","Data":"2e485bbf244c2eee2edddbce5118e3f678745a96e30d7346f11f5ad5b42d573b"} Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.925390 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae1255e2-6e7d-43d0-ab12-2af33964e2f4","Type":"ContainerDied","Data":"e21f3ffcdceacd7569d61433c93fcf96498efdc5108290f59f34bdd7729e73b0"} Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.925400 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae1255e2-6e7d-43d0-ab12-2af33964e2f4","Type":"ContainerDied","Data":"f05659e5fb73e35aa36f0b35a4df3bb3dcc03667b897b7e4af4fb48f282965b6"} Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.925408 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae1255e2-6e7d-43d0-ab12-2af33964e2f4","Type":"ContainerDied","Data":"58e9a1b30a4695ba81172383f9bc66a96743a5fcf919e2bb7a3a8df0b32b72a5"} Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.925252 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=3.18844598 podStartE2EDuration="3.925230306s" podCreationTimestamp="2026-03-13 09:39:04 +0000 UTC" firstStartedPulling="2026-03-13 09:39:05.950827269 +0000 UTC m=+1586.700741946" lastFinishedPulling="2026-03-13 09:39:06.687611595 +0000 UTC m=+1587.437526272" observedRunningTime="2026-03-13 09:39:07.919668795 +0000 UTC m=+1588.669583482" watchObservedRunningTime="2026-03-13 09:39:07.925230306 +0000 UTC m=+1588.675144983" Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.925351 4930 generic.go:334] "Generic (PLEG): container finished" podID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerID="58e9a1b30a4695ba81172383f9bc66a96743a5fcf919e2bb7a3a8df0b32b72a5" exitCode=0 Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.941877 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"79701ff1-8a11-4f24-9ab6-21f037545d5f","Type":"ContainerStarted","Data":"2bef53636a65aa4df31016d5e4a9dbca79a002bb165bc0e482564a6814cfcc43"} Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.943568 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 09:39:07 crc kubenswrapper[4930]: I0313 09:39:07.976942 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.046329481 podStartE2EDuration="3.976917768s" podCreationTimestamp="2026-03-13 09:39:04 +0000 UTC" firstStartedPulling="2026-03-13 09:39:06.045544493 +0000 UTC m=+1586.795459170" lastFinishedPulling="2026-03-13 09:39:06.97613278 +0000 UTC m=+1587.726047457" observedRunningTime="2026-03-13 09:39:07.9691185 +0000 UTC m=+1588.719033187" watchObservedRunningTime="2026-03-13 09:39:07.976917768 +0000 UTC m=+1588.726832445" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.564993 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.640252 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkg98\" (UniqueName: \"kubernetes.io/projected/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-kube-api-access-kkg98\") pod \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.640610 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-run-httpd\") pod \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.640700 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-sg-core-conf-yaml\") pod \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.640895 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-scripts\") pod \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.640992 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-config-data\") pod \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.641101 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-log-httpd\") pod \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.641218 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-combined-ca-bundle\") pod \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\" (UID: \"ae1255e2-6e7d-43d0-ab12-2af33964e2f4\") " Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.640922 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "ae1255e2-6e7d-43d0-ab12-2af33964e2f4" (UID: "ae1255e2-6e7d-43d0-ab12-2af33964e2f4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.643336 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "ae1255e2-6e7d-43d0-ab12-2af33964e2f4" (UID: "ae1255e2-6e7d-43d0-ab12-2af33964e2f4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.646971 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-kube-api-access-kkg98" (OuterVolumeSpecName: "kube-api-access-kkg98") pod "ae1255e2-6e7d-43d0-ab12-2af33964e2f4" (UID: "ae1255e2-6e7d-43d0-ab12-2af33964e2f4"). InnerVolumeSpecName "kube-api-access-kkg98". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.670984 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-scripts" (OuterVolumeSpecName: "scripts") pod "ae1255e2-6e7d-43d0-ab12-2af33964e2f4" (UID: "ae1255e2-6e7d-43d0-ab12-2af33964e2f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.700273 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "ae1255e2-6e7d-43d0-ab12-2af33964e2f4" (UID: "ae1255e2-6e7d-43d0-ab12-2af33964e2f4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.744010 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.744059 4930 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.744092 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkg98\" (UniqueName: \"kubernetes.io/projected/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-kube-api-access-kkg98\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.744104 4930 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.744117 4930 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.757407 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae1255e2-6e7d-43d0-ab12-2af33964e2f4" (UID: "ae1255e2-6e7d-43d0-ab12-2af33964e2f4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.793428 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-config-data" (OuterVolumeSpecName: "config-data") pod "ae1255e2-6e7d-43d0-ab12-2af33964e2f4" (UID: "ae1255e2-6e7d-43d0-ab12-2af33964e2f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.848729 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.849052 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae1255e2-6e7d-43d0-ab12-2af33964e2f4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.967083 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.967077 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ae1255e2-6e7d-43d0-ab12-2af33964e2f4","Type":"ContainerDied","Data":"4165ef7f0faa1d759b6a7a390da07b24f88711c3df60cb5c5e9d1592f6ef6a16"} Mar 13 09:39:08 crc kubenswrapper[4930]: I0313 09:39:08.967257 4930 scope.go:117] "RemoveContainer" containerID="2e485bbf244c2eee2edddbce5118e3f678745a96e30d7346f11f5ad5b42d573b" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.007103 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.007849 4930 scope.go:117] "RemoveContainer" containerID="e21f3ffcdceacd7569d61433c93fcf96498efdc5108290f59f34bdd7729e73b0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.021835 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.036773 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:39:09 crc kubenswrapper[4930]: E0313 09:39:09.037289 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="sg-core" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.037305 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="sg-core" Mar 13 09:39:09 crc kubenswrapper[4930]: E0313 09:39:09.037318 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="proxy-httpd" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.037324 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="proxy-httpd" Mar 13 09:39:09 crc kubenswrapper[4930]: E0313 09:39:09.037353 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="ceilometer-central-agent" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.037359 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="ceilometer-central-agent" Mar 13 09:39:09 crc kubenswrapper[4930]: E0313 09:39:09.037386 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="ceilometer-notification-agent" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.037392 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="ceilometer-notification-agent" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.037619 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="proxy-httpd" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.037634 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="sg-core" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.037655 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="ceilometer-notification-agent" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.037667 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" containerName="ceilometer-central-agent" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.039772 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.043180 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.043407 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.043574 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.046483 4930 scope.go:117] "RemoveContainer" containerID="f05659e5fb73e35aa36f0b35a4df3bb3dcc03667b897b7e4af4fb48f282965b6" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.076406 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.082734 4930 scope.go:117] "RemoveContainer" containerID="58e9a1b30a4695ba81172383f9bc66a96743a5fcf919e2bb7a3a8df0b32b72a5" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.160868 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.160970 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.161007 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.161136 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-config-data\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.161347 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbk7k\" (UniqueName: \"kubernetes.io/projected/0372477d-d529-4348-add4-b8d5f63c1acc-kube-api-access-lbk7k\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.161776 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0372477d-d529-4348-add4-b8d5f63c1acc-run-httpd\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.161848 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-scripts\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.162071 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0372477d-d529-4348-add4-b8d5f63c1acc-log-httpd\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.264374 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0372477d-d529-4348-add4-b8d5f63c1acc-log-httpd\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.264511 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.264592 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.264628 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.264671 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-config-data\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.264736 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbk7k\" (UniqueName: \"kubernetes.io/projected/0372477d-d529-4348-add4-b8d5f63c1acc-kube-api-access-lbk7k\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.264801 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0372477d-d529-4348-add4-b8d5f63c1acc-run-httpd\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.264838 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-scripts\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.264944 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0372477d-d529-4348-add4-b8d5f63c1acc-log-httpd\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.266299 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0372477d-d529-4348-add4-b8d5f63c1acc-run-httpd\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.269346 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-config-data\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.269997 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.270113 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.270225 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-scripts\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.270251 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.282925 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbk7k\" (UniqueName: \"kubernetes.io/projected/0372477d-d529-4348-add4-b8d5f63c1acc-kube-api-access-lbk7k\") pod \"ceilometer-0\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.362405 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:39:09 crc kubenswrapper[4930]: I0313 09:39:09.852581 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:39:09 crc kubenswrapper[4930]: W0313 09:39:09.866626 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0372477d_d529_4348_add4_b8d5f63c1acc.slice/crio-40d8f484c6f55ff7b491cd7bae9ae4e7b841a5cc321f77d6be60f3dde561f207 WatchSource:0}: Error finding container 40d8f484c6f55ff7b491cd7bae9ae4e7b841a5cc321f77d6be60f3dde561f207: Status 404 returned error can't find the container with id 40d8f484c6f55ff7b491cd7bae9ae4e7b841a5cc321f77d6be60f3dde561f207 Mar 13 09:39:10 crc kubenswrapper[4930]: I0313 09:39:09.994293 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae1255e2-6e7d-43d0-ab12-2af33964e2f4" path="/var/lib/kubelet/pods/ae1255e2-6e7d-43d0-ab12-2af33964e2f4/volumes" Mar 13 09:39:10 crc kubenswrapper[4930]: I0313 09:39:09.996107 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0372477d-d529-4348-add4-b8d5f63c1acc","Type":"ContainerStarted","Data":"40d8f484c6f55ff7b491cd7bae9ae4e7b841a5cc321f77d6be60f3dde561f207"} Mar 13 09:39:11 crc kubenswrapper[4930]: I0313 09:39:11.015304 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0372477d-d529-4348-add4-b8d5f63c1acc","Type":"ContainerStarted","Data":"fc8185bc3d1a5927e832d10488bf14333862645bf76c4c34cb5c2261438451de"} Mar 13 09:39:12 crc kubenswrapper[4930]: I0313 09:39:12.035344 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0372477d-d529-4348-add4-b8d5f63c1acc","Type":"ContainerStarted","Data":"2bebd1bb9825faee9e9878ce84f5b67960760530b6a0f9ffab50e0626d2f8a00"} Mar 13 09:39:12 crc kubenswrapper[4930]: I0313 09:39:12.971738 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:39:12 crc kubenswrapper[4930]: E0313 09:39:12.972693 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:39:13 crc kubenswrapper[4930]: I0313 09:39:13.092638 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0372477d-d529-4348-add4-b8d5f63c1acc","Type":"ContainerStarted","Data":"41be0d2fb5d85853126a1ed22098f3d3d97f386da77abab1a5e1d6d1112dd42c"} Mar 13 09:39:13 crc kubenswrapper[4930]: I0313 09:39:13.858210 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-p4nk9"] Mar 13 09:39:13 crc kubenswrapper[4930]: I0313 09:39:13.869729 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-p4nk9"] Mar 13 09:39:13 crc kubenswrapper[4930]: I0313 09:39:13.992096 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae61cc80-b8f2-401c-87db-5728d566b288" path="/var/lib/kubelet/pods/ae61cc80-b8f2-401c-87db-5728d566b288/volumes" Mar 13 09:39:13 crc kubenswrapper[4930]: I0313 09:39:13.996250 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-6sjqb"] Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.005635 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6sjqb" Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.017445 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-6sjqb"] Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.123149 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gm7g\" (UniqueName: \"kubernetes.io/projected/94958e65-6d03-4346-b6c0-47e3de576961-kube-api-access-7gm7g\") pod \"heat-db-sync-6sjqb\" (UID: \"94958e65-6d03-4346-b6c0-47e3de576961\") " pod="openstack/heat-db-sync-6sjqb" Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.124694 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94958e65-6d03-4346-b6c0-47e3de576961-config-data\") pod \"heat-db-sync-6sjqb\" (UID: \"94958e65-6d03-4346-b6c0-47e3de576961\") " pod="openstack/heat-db-sync-6sjqb" Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.125470 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94958e65-6d03-4346-b6c0-47e3de576961-combined-ca-bundle\") pod \"heat-db-sync-6sjqb\" (UID: \"94958e65-6d03-4346-b6c0-47e3de576961\") " pod="openstack/heat-db-sync-6sjqb" Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.231471 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gm7g\" (UniqueName: \"kubernetes.io/projected/94958e65-6d03-4346-b6c0-47e3de576961-kube-api-access-7gm7g\") pod \"heat-db-sync-6sjqb\" (UID: \"94958e65-6d03-4346-b6c0-47e3de576961\") " pod="openstack/heat-db-sync-6sjqb" Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.232018 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94958e65-6d03-4346-b6c0-47e3de576961-config-data\") pod \"heat-db-sync-6sjqb\" (UID: \"94958e65-6d03-4346-b6c0-47e3de576961\") " pod="openstack/heat-db-sync-6sjqb" Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.232063 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94958e65-6d03-4346-b6c0-47e3de576961-combined-ca-bundle\") pod \"heat-db-sync-6sjqb\" (UID: \"94958e65-6d03-4346-b6c0-47e3de576961\") " pod="openstack/heat-db-sync-6sjqb" Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.241263 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94958e65-6d03-4346-b6c0-47e3de576961-config-data\") pod \"heat-db-sync-6sjqb\" (UID: \"94958e65-6d03-4346-b6c0-47e3de576961\") " pod="openstack/heat-db-sync-6sjqb" Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.241764 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94958e65-6d03-4346-b6c0-47e3de576961-combined-ca-bundle\") pod \"heat-db-sync-6sjqb\" (UID: \"94958e65-6d03-4346-b6c0-47e3de576961\") " pod="openstack/heat-db-sync-6sjqb" Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.255329 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gm7g\" (UniqueName: \"kubernetes.io/projected/94958e65-6d03-4346-b6c0-47e3de576961-kube-api-access-7gm7g\") pod \"heat-db-sync-6sjqb\" (UID: \"94958e65-6d03-4346-b6c0-47e3de576961\") " pod="openstack/heat-db-sync-6sjqb" Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.370393 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6sjqb" Mar 13 09:39:14 crc kubenswrapper[4930]: I0313 09:39:14.928439 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-6sjqb"] Mar 13 09:39:14 crc kubenswrapper[4930]: W0313 09:39:14.934577 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod94958e65_6d03_4346_b6c0_47e3de576961.slice/crio-b713f4a7271e88352615848e0355fa77409615c8dc2f9f940c3702e1ee0fc1bb WatchSource:0}: Error finding container b713f4a7271e88352615848e0355fa77409615c8dc2f9f940c3702e1ee0fc1bb: Status 404 returned error can't find the container with id b713f4a7271e88352615848e0355fa77409615c8dc2f9f940c3702e1ee0fc1bb Mar 13 09:39:15 crc kubenswrapper[4930]: I0313 09:39:15.115480 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0372477d-d529-4348-add4-b8d5f63c1acc","Type":"ContainerStarted","Data":"3713963a3bc527583bcc18c3102e8088363239020ef2b832f0682be5a198bd65"} Mar 13 09:39:15 crc kubenswrapper[4930]: I0313 09:39:15.115695 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 09:39:15 crc kubenswrapper[4930]: I0313 09:39:15.117202 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6sjqb" event={"ID":"94958e65-6d03-4346-b6c0-47e3de576961","Type":"ContainerStarted","Data":"b713f4a7271e88352615848e0355fa77409615c8dc2f9f940c3702e1ee0fc1bb"} Mar 13 09:39:15 crc kubenswrapper[4930]: I0313 09:39:15.150524 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.840254935 podStartE2EDuration="6.150503145s" podCreationTimestamp="2026-03-13 09:39:09 +0000 UTC" firstStartedPulling="2026-03-13 09:39:09.869136029 +0000 UTC m=+1590.619050706" lastFinishedPulling="2026-03-13 09:39:14.179384239 +0000 UTC m=+1594.929298916" observedRunningTime="2026-03-13 09:39:15.133537134 +0000 UTC m=+1595.883451831" watchObservedRunningTime="2026-03-13 09:39:15.150503145 +0000 UTC m=+1595.900417822" Mar 13 09:39:15 crc kubenswrapper[4930]: I0313 09:39:15.451972 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 09:39:16 crc kubenswrapper[4930]: I0313 09:39:16.265503 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 13 09:39:17 crc kubenswrapper[4930]: I0313 09:39:17.521858 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 09:39:18 crc kubenswrapper[4930]: I0313 09:39:18.200540 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:39:18 crc kubenswrapper[4930]: I0313 09:39:18.201063 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="ceilometer-central-agent" containerID="cri-o://fc8185bc3d1a5927e832d10488bf14333862645bf76c4c34cb5c2261438451de" gracePeriod=30 Mar 13 09:39:18 crc kubenswrapper[4930]: I0313 09:39:18.201177 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="ceilometer-notification-agent" containerID="cri-o://2bebd1bb9825faee9e9878ce84f5b67960760530b6a0f9ffab50e0626d2f8a00" gracePeriod=30 Mar 13 09:39:18 crc kubenswrapper[4930]: I0313 09:39:18.201141 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="sg-core" containerID="cri-o://41be0d2fb5d85853126a1ed22098f3d3d97f386da77abab1a5e1d6d1112dd42c" gracePeriod=30 Mar 13 09:39:18 crc kubenswrapper[4930]: I0313 09:39:18.201091 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="proxy-httpd" containerID="cri-o://3713963a3bc527583bcc18c3102e8088363239020ef2b832f0682be5a198bd65" gracePeriod=30 Mar 13 09:39:19 crc kubenswrapper[4930]: I0313 09:39:19.183912 4930 generic.go:334] "Generic (PLEG): container finished" podID="0372477d-d529-4348-add4-b8d5f63c1acc" containerID="3713963a3bc527583bcc18c3102e8088363239020ef2b832f0682be5a198bd65" exitCode=0 Mar 13 09:39:19 crc kubenswrapper[4930]: I0313 09:39:19.184273 4930 generic.go:334] "Generic (PLEG): container finished" podID="0372477d-d529-4348-add4-b8d5f63c1acc" containerID="41be0d2fb5d85853126a1ed22098f3d3d97f386da77abab1a5e1d6d1112dd42c" exitCode=2 Mar 13 09:39:19 crc kubenswrapper[4930]: I0313 09:39:19.184284 4930 generic.go:334] "Generic (PLEG): container finished" podID="0372477d-d529-4348-add4-b8d5f63c1acc" containerID="2bebd1bb9825faee9e9878ce84f5b67960760530b6a0f9ffab50e0626d2f8a00" exitCode=0 Mar 13 09:39:19 crc kubenswrapper[4930]: I0313 09:39:19.183946 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0372477d-d529-4348-add4-b8d5f63c1acc","Type":"ContainerDied","Data":"3713963a3bc527583bcc18c3102e8088363239020ef2b832f0682be5a198bd65"} Mar 13 09:39:19 crc kubenswrapper[4930]: I0313 09:39:19.184326 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0372477d-d529-4348-add4-b8d5f63c1acc","Type":"ContainerDied","Data":"41be0d2fb5d85853126a1ed22098f3d3d97f386da77abab1a5e1d6d1112dd42c"} Mar 13 09:39:19 crc kubenswrapper[4930]: I0313 09:39:19.184342 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0372477d-d529-4348-add4-b8d5f63c1acc","Type":"ContainerDied","Data":"2bebd1bb9825faee9e9878ce84f5b67960760530b6a0f9ffab50e0626d2f8a00"} Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.214161 4930 generic.go:334] "Generic (PLEG): container finished" podID="0372477d-d529-4348-add4-b8d5f63c1acc" containerID="fc8185bc3d1a5927e832d10488bf14333862645bf76c4c34cb5c2261438451de" exitCode=0 Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.214233 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0372477d-d529-4348-add4-b8d5f63c1acc","Type":"ContainerDied","Data":"fc8185bc3d1a5927e832d10488bf14333862645bf76c4c34cb5c2261438451de"} Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.375591 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.439818 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-sg-core-conf-yaml\") pod \"0372477d-d529-4348-add4-b8d5f63c1acc\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.439927 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-config-data\") pod \"0372477d-d529-4348-add4-b8d5f63c1acc\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.439963 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-combined-ca-bundle\") pod \"0372477d-d529-4348-add4-b8d5f63c1acc\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.439992 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-ceilometer-tls-certs\") pod \"0372477d-d529-4348-add4-b8d5f63c1acc\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.440038 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0372477d-d529-4348-add4-b8d5f63c1acc-log-httpd\") pod \"0372477d-d529-4348-add4-b8d5f63c1acc\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.440087 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-scripts\") pod \"0372477d-d529-4348-add4-b8d5f63c1acc\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.440125 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbk7k\" (UniqueName: \"kubernetes.io/projected/0372477d-d529-4348-add4-b8d5f63c1acc-kube-api-access-lbk7k\") pod \"0372477d-d529-4348-add4-b8d5f63c1acc\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.440191 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0372477d-d529-4348-add4-b8d5f63c1acc-run-httpd\") pod \"0372477d-d529-4348-add4-b8d5f63c1acc\" (UID: \"0372477d-d529-4348-add4-b8d5f63c1acc\") " Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.442128 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0372477d-d529-4348-add4-b8d5f63c1acc-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "0372477d-d529-4348-add4-b8d5f63c1acc" (UID: "0372477d-d529-4348-add4-b8d5f63c1acc"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.450392 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0372477d-d529-4348-add4-b8d5f63c1acc-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "0372477d-d529-4348-add4-b8d5f63c1acc" (UID: "0372477d-d529-4348-add4-b8d5f63c1acc"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.456604 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0372477d-d529-4348-add4-b8d5f63c1acc-kube-api-access-lbk7k" (OuterVolumeSpecName: "kube-api-access-lbk7k") pod "0372477d-d529-4348-add4-b8d5f63c1acc" (UID: "0372477d-d529-4348-add4-b8d5f63c1acc"). InnerVolumeSpecName "kube-api-access-lbk7k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.456930 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-scripts" (OuterVolumeSpecName: "scripts") pod "0372477d-d529-4348-add4-b8d5f63c1acc" (UID: "0372477d-d529-4348-add4-b8d5f63c1acc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.543915 4930 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0372477d-d529-4348-add4-b8d5f63c1acc-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.544290 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.544383 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbk7k\" (UniqueName: \"kubernetes.io/projected/0372477d-d529-4348-add4-b8d5f63c1acc-kube-api-access-lbk7k\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.544464 4930 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/0372477d-d529-4348-add4-b8d5f63c1acc-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.544706 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "0372477d-d529-4348-add4-b8d5f63c1acc" (UID: "0372477d-d529-4348-add4-b8d5f63c1acc"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.582984 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "0372477d-d529-4348-add4-b8d5f63c1acc" (UID: "0372477d-d529-4348-add4-b8d5f63c1acc"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.653864 4930 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.653912 4930 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.675738 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0372477d-d529-4348-add4-b8d5f63c1acc" (UID: "0372477d-d529-4348-add4-b8d5f63c1acc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.696208 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-config-data" (OuterVolumeSpecName: "config-data") pod "0372477d-d529-4348-add4-b8d5f63c1acc" (UID: "0372477d-d529-4348-add4-b8d5f63c1acc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.756089 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:20 crc kubenswrapper[4930]: I0313 09:39:20.756120 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0372477d-d529-4348-add4-b8d5f63c1acc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.232711 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"0372477d-d529-4348-add4-b8d5f63c1acc","Type":"ContainerDied","Data":"40d8f484c6f55ff7b491cd7bae9ae4e7b841a5cc321f77d6be60f3dde561f207"} Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.233036 4930 scope.go:117] "RemoveContainer" containerID="3713963a3bc527583bcc18c3102e8088363239020ef2b832f0682be5a198bd65" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.232850 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.334613 4930 scope.go:117] "RemoveContainer" containerID="41be0d2fb5d85853126a1ed22098f3d3d97f386da77abab1a5e1d6d1112dd42c" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.358299 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.363656 4930 scope.go:117] "RemoveContainer" containerID="2bebd1bb9825faee9e9878ce84f5b67960760530b6a0f9ffab50e0626d2f8a00" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.378228 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.579427 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:39:21 crc kubenswrapper[4930]: E0313 09:39:21.581738 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="sg-core" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.581776 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="sg-core" Mar 13 09:39:21 crc kubenswrapper[4930]: E0313 09:39:21.581839 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="ceilometer-central-agent" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.581849 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="ceilometer-central-agent" Mar 13 09:39:21 crc kubenswrapper[4930]: E0313 09:39:21.581860 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="proxy-httpd" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.581871 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="proxy-httpd" Mar 13 09:39:21 crc kubenswrapper[4930]: E0313 09:39:21.581884 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="ceilometer-notification-agent" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.581891 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="ceilometer-notification-agent" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.582173 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="ceilometer-central-agent" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.582204 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="proxy-httpd" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.582223 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="sg-core" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.582235 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" containerName="ceilometer-notification-agent" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.586058 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.601824 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.601906 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.602085 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.631367 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.645218 4930 scope.go:117] "RemoveContainer" containerID="fc8185bc3d1a5927e832d10488bf14333862645bf76c4c34cb5c2261438451de" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.685877 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4640978-a913-419b-9faa-8b230f5e51f2-run-httpd\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.686040 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-scripts\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.686077 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.686161 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snkkd\" (UniqueName: \"kubernetes.io/projected/d4640978-a913-419b-9faa-8b230f5e51f2-kube-api-access-snkkd\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.686202 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.686289 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-config-data\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.686328 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4640978-a913-419b-9faa-8b230f5e51f2-log-httpd\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.686365 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.794207 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snkkd\" (UniqueName: \"kubernetes.io/projected/d4640978-a913-419b-9faa-8b230f5e51f2-kube-api-access-snkkd\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.794263 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.794347 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-config-data\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.794375 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4640978-a913-419b-9faa-8b230f5e51f2-log-httpd\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.794403 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.796243 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4640978-a913-419b-9faa-8b230f5e51f2-run-httpd\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.796414 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-scripts\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.796462 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.796929 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4640978-a913-419b-9faa-8b230f5e51f2-log-httpd\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.797161 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4640978-a913-419b-9faa-8b230f5e51f2-run-httpd\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.803001 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-scripts\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.804489 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-config-data\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.815396 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.817605 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.818886 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.828808 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snkkd\" (UniqueName: \"kubernetes.io/projected/d4640978-a913-419b-9faa-8b230f5e51f2-kube-api-access-snkkd\") pod \"ceilometer-0\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.926127 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 09:39:21 crc kubenswrapper[4930]: I0313 09:39:21.987901 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0372477d-d529-4348-add4-b8d5f63c1acc" path="/var/lib/kubelet/pods/0372477d-d529-4348-add4-b8d5f63c1acc/volumes" Mar 13 09:39:22 crc kubenswrapper[4930]: I0313 09:39:22.350812 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-2" podUID="a6623f09-f961-411a-9323-8cbf953c7148" containerName="rabbitmq" containerID="cri-o://b379c2c9b1a4f5934e943fc74b3b6eee20c34346290ab3b05975896af875e676" gracePeriod=604794 Mar 13 09:39:22 crc kubenswrapper[4930]: I0313 09:39:22.550250 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 09:39:22 crc kubenswrapper[4930]: W0313 09:39:22.554624 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd4640978_a913_419b_9faa_8b230f5e51f2.slice/crio-2f4dbf9dcdf28ac45b6b206d225da1af5e1cfdfabc95f44458f1e62ed25c5cc8 WatchSource:0}: Error finding container 2f4dbf9dcdf28ac45b6b206d225da1af5e1cfdfabc95f44458f1e62ed25c5cc8: Status 404 returned error can't find the container with id 2f4dbf9dcdf28ac45b6b206d225da1af5e1cfdfabc95f44458f1e62ed25c5cc8 Mar 13 09:39:22 crc kubenswrapper[4930]: I0313 09:39:22.795406 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" containerName="rabbitmq" containerID="cri-o://5d93f52f3a7b4038546d2005345570ab31cd4165cbad2ec0b14428854e754b01" gracePeriod=604795 Mar 13 09:39:23 crc kubenswrapper[4930]: I0313 09:39:23.120129 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="a6623f09-f961-411a-9323-8cbf953c7148" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.136:5671: connect: connection refused" Mar 13 09:39:23 crc kubenswrapper[4930]: I0313 09:39:23.302196 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4640978-a913-419b-9faa-8b230f5e51f2","Type":"ContainerStarted","Data":"2f4dbf9dcdf28ac45b6b206d225da1af5e1cfdfabc95f44458f1e62ed25c5cc8"} Mar 13 09:39:23 crc kubenswrapper[4930]: I0313 09:39:23.985047 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:39:23 crc kubenswrapper[4930]: E0313 09:39:23.986031 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:39:29 crc kubenswrapper[4930]: I0313 09:39:29.400984 4930 generic.go:334] "Generic (PLEG): container finished" podID="dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" containerID="5d93f52f3a7b4038546d2005345570ab31cd4165cbad2ec0b14428854e754b01" exitCode=0 Mar 13 09:39:29 crc kubenswrapper[4930]: I0313 09:39:29.401066 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354","Type":"ContainerDied","Data":"5d93f52f3a7b4038546d2005345570ab31cd4165cbad2ec0b14428854e754b01"} Mar 13 09:39:29 crc kubenswrapper[4930]: I0313 09:39:29.405447 4930 generic.go:334] "Generic (PLEG): container finished" podID="a6623f09-f961-411a-9323-8cbf953c7148" containerID="b379c2c9b1a4f5934e943fc74b3b6eee20c34346290ab3b05975896af875e676" exitCode=0 Mar 13 09:39:29 crc kubenswrapper[4930]: I0313 09:39:29.405541 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a6623f09-f961-411a-9323-8cbf953c7148","Type":"ContainerDied","Data":"b379c2c9b1a4f5934e943fc74b3b6eee20c34346290ab3b05975896af875e676"} Mar 13 09:39:32 crc kubenswrapper[4930]: I0313 09:39:32.591310 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.134:5671: connect: connection refused" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.736600 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-n25jr"] Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.738987 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.741646 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.751458 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-n25jr"] Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.837532 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.837601 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84dtz\" (UniqueName: \"kubernetes.io/projected/0a90a3cb-c199-4c04-a702-7cc29427a444-kube-api-access-84dtz\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.837671 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-config\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.837730 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.837872 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.837969 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.838006 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.940050 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.940114 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.940269 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.940378 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84dtz\" (UniqueName: \"kubernetes.io/projected/0a90a3cb-c199-4c04-a702-7cc29427a444-kube-api-access-84dtz\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.940399 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-config\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.940423 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.941109 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-ovsdbserver-sb\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.941157 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.941215 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-dns-swift-storage-0\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.941540 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-config\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.941653 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-dns-svc\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.941812 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-openstack-edpm-ipam\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.943232 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-ovsdbserver-nb\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:33 crc kubenswrapper[4930]: I0313 09:39:33.962185 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84dtz\" (UniqueName: \"kubernetes.io/projected/0a90a3cb-c199-4c04-a702-7cc29427a444-kube-api-access-84dtz\") pod \"dnsmasq-dns-5b75489c6f-n25jr\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:34 crc kubenswrapper[4930]: I0313 09:39:34.062958 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:34 crc kubenswrapper[4930]: I0313 09:39:34.981700 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.076488 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-config-data\") pod \"a6623f09-f961-411a-9323-8cbf953c7148\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.076677 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-tls\") pod \"a6623f09-f961-411a-9323-8cbf953c7148\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.076707 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-confd\") pod \"a6623f09-f961-411a-9323-8cbf953c7148\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.076722 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-plugins-conf\") pod \"a6623f09-f961-411a-9323-8cbf953c7148\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.082246 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "a6623f09-f961-411a-9323-8cbf953c7148" (UID: "a6623f09-f961-411a-9323-8cbf953c7148"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.106516 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "a6623f09-f961-411a-9323-8cbf953c7148" (UID: "a6623f09-f961-411a-9323-8cbf953c7148"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.117072 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\") pod \"a6623f09-f961-411a-9323-8cbf953c7148\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.117187 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-plugins\") pod \"a6623f09-f961-411a-9323-8cbf953c7148\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.117226 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6623f09-f961-411a-9323-8cbf953c7148-pod-info\") pod \"a6623f09-f961-411a-9323-8cbf953c7148\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.117334 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6623f09-f961-411a-9323-8cbf953c7148-erlang-cookie-secret\") pod \"a6623f09-f961-411a-9323-8cbf953c7148\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.117365 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4976\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-kube-api-access-j4976\") pod \"a6623f09-f961-411a-9323-8cbf953c7148\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.117392 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-erlang-cookie\") pod \"a6623f09-f961-411a-9323-8cbf953c7148\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.117419 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-server-conf\") pod \"a6623f09-f961-411a-9323-8cbf953c7148\" (UID: \"a6623f09-f961-411a-9323-8cbf953c7148\") " Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.118462 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.118474 4930 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.124789 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "a6623f09-f961-411a-9323-8cbf953c7148" (UID: "a6623f09-f961-411a-9323-8cbf953c7148"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.126104 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-kube-api-access-j4976" (OuterVolumeSpecName: "kube-api-access-j4976") pod "a6623f09-f961-411a-9323-8cbf953c7148" (UID: "a6623f09-f961-411a-9323-8cbf953c7148"). InnerVolumeSpecName "kube-api-access-j4976". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.136530 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/a6623f09-f961-411a-9323-8cbf953c7148-pod-info" (OuterVolumeSpecName: "pod-info") pod "a6623f09-f961-411a-9323-8cbf953c7148" (UID: "a6623f09-f961-411a-9323-8cbf953c7148"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.151947 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "a6623f09-f961-411a-9323-8cbf953c7148" (UID: "a6623f09-f961-411a-9323-8cbf953c7148"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.152986 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a6623f09-f961-411a-9323-8cbf953c7148-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "a6623f09-f961-411a-9323-8cbf953c7148" (UID: "a6623f09-f961-411a-9323-8cbf953c7148"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.183076 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-config-data" (OuterVolumeSpecName: "config-data") pod "a6623f09-f961-411a-9323-8cbf953c7148" (UID: "a6623f09-f961-411a-9323-8cbf953c7148"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.191992 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5" (OuterVolumeSpecName: "persistence") pod "a6623f09-f961-411a-9323-8cbf953c7148" (UID: "a6623f09-f961-411a-9323-8cbf953c7148"). InnerVolumeSpecName "pvc-e80c2007-88e5-41ef-a376-d647dd6202b5". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.217863 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-server-conf" (OuterVolumeSpecName: "server-conf") pod "a6623f09-f961-411a-9323-8cbf953c7148" (UID: "a6623f09-f961-411a-9323-8cbf953c7148"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.221108 4930 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\") on node \"crc\" " Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.221146 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.221159 4930 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/a6623f09-f961-411a-9323-8cbf953c7148-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.221171 4930 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/a6623f09-f961-411a-9323-8cbf953c7148-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.221180 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4976\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-kube-api-access-j4976\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.221190 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.221198 4930 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.221206 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/a6623f09-f961-411a-9323-8cbf953c7148-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.275867 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "a6623f09-f961-411a-9323-8cbf953c7148" (UID: "a6623f09-f961-411a-9323-8cbf953c7148"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.281482 4930 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.281623 4930 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-e80c2007-88e5-41ef-a376-d647dd6202b5" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5") on node "crc" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.324948 4930 reconciler_common.go:293] "Volume detached for volume \"pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.325189 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/a6623f09-f961-411a-9323-8cbf953c7148-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.535593 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"a6623f09-f961-411a-9323-8cbf953c7148","Type":"ContainerDied","Data":"2cedef14877fa4198d43afb9b9dc678dbb52d67071539ef6175ee15d5d9be9a1"} Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.535666 4930 scope.go:117] "RemoveContainer" containerID="b379c2c9b1a4f5934e943fc74b3b6eee20c34346290ab3b05975896af875e676" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.535746 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.625923 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.646512 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.667105 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-2"] Mar 13 09:39:35 crc kubenswrapper[4930]: E0313 09:39:35.667785 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6623f09-f961-411a-9323-8cbf953c7148" containerName="setup-container" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.667808 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6623f09-f961-411a-9323-8cbf953c7148" containerName="setup-container" Mar 13 09:39:35 crc kubenswrapper[4930]: E0313 09:39:35.667840 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6623f09-f961-411a-9323-8cbf953c7148" containerName="rabbitmq" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.667850 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6623f09-f961-411a-9323-8cbf953c7148" containerName="rabbitmq" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.668139 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6623f09-f961-411a-9323-8cbf953c7148" containerName="rabbitmq" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.669936 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.745389 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc0f868a-2947-4e5a-a502-7ece4237551c-pod-info\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.745459 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc0f868a-2947-4e5a-a502-7ece4237551c-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.745580 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc0f868a-2947-4e5a-a502-7ece4237551c-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.745667 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc0f868a-2947-4e5a-a502-7ece4237551c-config-data\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.745686 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc0f868a-2947-4e5a-a502-7ece4237551c-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.745711 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24s77\" (UniqueName: \"kubernetes.io/projected/bc0f868a-2947-4e5a-a502-7ece4237551c-kube-api-access-24s77\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.745841 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc0f868a-2947-4e5a-a502-7ece4237551c-server-conf\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.745888 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc0f868a-2947-4e5a-a502-7ece4237551c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.745980 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.746112 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc0f868a-2947-4e5a-a502-7ece4237551c-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.746258 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc0f868a-2947-4e5a-a502-7ece4237551c-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.768214 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.850989 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc0f868a-2947-4e5a-a502-7ece4237551c-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.851118 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc0f868a-2947-4e5a-a502-7ece4237551c-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.851151 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc0f868a-2947-4e5a-a502-7ece4237551c-pod-info\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.851203 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc0f868a-2947-4e5a-a502-7ece4237551c-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.851247 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc0f868a-2947-4e5a-a502-7ece4237551c-config-data\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.851269 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc0f868a-2947-4e5a-a502-7ece4237551c-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.851293 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24s77\" (UniqueName: \"kubernetes.io/projected/bc0f868a-2947-4e5a-a502-7ece4237551c-kube-api-access-24s77\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.851369 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc0f868a-2947-4e5a-a502-7ece4237551c-server-conf\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.851408 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc0f868a-2947-4e5a-a502-7ece4237551c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.851497 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.851580 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc0f868a-2947-4e5a-a502-7ece4237551c-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.854973 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/bc0f868a-2947-4e5a-a502-7ece4237551c-erlang-cookie-secret\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.856525 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/bc0f868a-2947-4e5a-a502-7ece4237551c-pod-info\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.857816 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/bc0f868a-2947-4e5a-a502-7ece4237551c-rabbitmq-plugins\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.858548 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.858598 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/e16cd85ffa615f778b152995784bb8923bd8ede16bc037cea68a4cfdb6f64ffd/globalmount\"" pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.859208 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/bc0f868a-2947-4e5a-a502-7ece4237551c-rabbitmq-confd\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.861376 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/bc0f868a-2947-4e5a-a502-7ece4237551c-plugins-conf\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.861640 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/bc0f868a-2947-4e5a-a502-7ece4237551c-server-conf\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.862061 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bc0f868a-2947-4e5a-a502-7ece4237551c-config-data\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.862350 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/bc0f868a-2947-4e5a-a502-7ece4237551c-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.863118 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/bc0f868a-2947-4e5a-a502-7ece4237551c-rabbitmq-tls\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.887071 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24s77\" (UniqueName: \"kubernetes.io/projected/bc0f868a-2947-4e5a-a502-7ece4237551c-kube-api-access-24s77\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.986880 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6623f09-f961-411a-9323-8cbf953c7148" path="/var/lib/kubelet/pods/a6623f09-f961-411a-9323-8cbf953c7148/volumes" Mar 13 09:39:35 crc kubenswrapper[4930]: I0313 09:39:35.999227 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-e80c2007-88e5-41ef-a376-d647dd6202b5\") pod \"rabbitmq-server-2\" (UID: \"bc0f868a-2947-4e5a-a502-7ece4237551c\") " pod="openstack/rabbitmq-server-2" Mar 13 09:39:36 crc kubenswrapper[4930]: I0313 09:39:36.039688 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-2" Mar 13 09:39:36 crc kubenswrapper[4930]: I0313 09:39:36.971086 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:39:36 crc kubenswrapper[4930]: E0313 09:39:36.971772 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:39:38 crc kubenswrapper[4930]: I0313 09:39:38.152991 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="a6623f09-f961-411a-9323-8cbf953c7148" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.136:5671: i/o timeout" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.622947 4930 scope.go:117] "RemoveContainer" containerID="96c39c7b929d20787c563592a28f65306475af49da6b2538643645d233ad69d4" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.668473 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354","Type":"ContainerDied","Data":"d9d4aecc2b7d2feebb05a1139df043bc7f059ac862bb780cb0ef82fafe45a341"} Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.668955 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d9d4aecc2b7d2feebb05a1139df043bc7f059ac862bb780cb0ef82fafe45a341" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.769690 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.802354 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-plugins-conf\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.802425 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65qcg\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-kube-api-access-65qcg\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.802477 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-tls\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.802508 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-erlang-cookie-secret\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.802531 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-config-data\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.802559 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-erlang-cookie\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.802581 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-confd\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.802604 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-pod-info\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.803594 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" (UID: "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.803819 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.803919 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-server-conf\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.803982 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-plugins\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") " Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.804119 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" (UID: "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.804855 4930 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.804877 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.814576 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" (UID: "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.815084 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-pod-info" (OuterVolumeSpecName: "pod-info") pod "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" (UID: "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.816155 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-kube-api-access-65qcg" (OuterVolumeSpecName: "kube-api-access-65qcg") pod "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" (UID: "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354"). InnerVolumeSpecName "kube-api-access-65qcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.816234 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" (UID: "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.873941 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-config-data" (OuterVolumeSpecName: "config-data") pod "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" (UID: "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.891132 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" (UID: "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.914564 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0" (OuterVolumeSpecName: "persistence") pod "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" (UID: "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354"). InnerVolumeSpecName "pvc-cb39e555-44ae-4231-8aab-8695177f64b0". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 09:39:46 crc kubenswrapper[4930]: E0313 09:39:46.915867 4930 reconciler_common.go:156] "operationExecutor.UnmountVolume failed (controllerAttachDetachEnabled true) for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") : UnmountVolume.NewUnmounter failed for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") : kubernetes.io/csi: unmounter failed to load volume data file [/var/lib/kubelet/pods/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354/volumes/kubernetes.io~csi/pvc-cb39e555-44ae-4231-8aab-8695177f64b0/mount]: kubernetes.io/csi: failed to open volume data file [/var/lib/kubelet/pods/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354/volumes/kubernetes.io~csi/pvc-cb39e555-44ae-4231-8aab-8695177f64b0/vol_data.json]: open /var/lib/kubelet/pods/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354/volumes/kubernetes.io~csi/pvc-cb39e555-44ae-4231-8aab-8695177f64b0/vol_data.json: no such file or directory" err="UnmountVolume.NewUnmounter failed for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") pod \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\" (UID: \"dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354\") : kubernetes.io/csi: unmounter failed to load volume data file [/var/lib/kubelet/pods/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354/volumes/kubernetes.io~csi/pvc-cb39e555-44ae-4231-8aab-8695177f64b0/mount]: kubernetes.io/csi: failed to open volume data file [/var/lib/kubelet/pods/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354/volumes/kubernetes.io~csi/pvc-cb39e555-44ae-4231-8aab-8695177f64b0/vol_data.json]: open /var/lib/kubelet/pods/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354/volumes/kubernetes.io~csi/pvc-cb39e555-44ae-4231-8aab-8695177f64b0/vol_data.json: no such file or directory" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.923385 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65qcg\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-kube-api-access-65qcg\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.923452 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.923473 4930 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.923486 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.923497 4930 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.923543 4930 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-cb39e555-44ae-4231-8aab-8695177f64b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") on node \"crc\" " Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.923561 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.978902 4930 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.979979 4930 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-cb39e555-44ae-4231-8aab-8695177f64b0" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0") on node "crc" Mar 13 09:39:46 crc kubenswrapper[4930]: I0313 09:39:46.987461 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-server-conf" (OuterVolumeSpecName: "server-conf") pod "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" (UID: "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.025972 4930 reconciler_common.go:293] "Volume detached for volume \"pvc-cb39e555-44ae-4231-8aab-8695177f64b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.026007 4930 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.057812 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" (UID: "dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.128198 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 09:39:47 crc kubenswrapper[4930]: E0313 09:39:47.237570 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Mar 13 09:39:47 crc kubenswrapper[4930]: E0313 09:39:47.237626 4930 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Mar 13 09:39:47 crc kubenswrapper[4930]: E0313 09:39:47.237776 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7gm7g,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-6sjqb_openstack(94958e65-6d03-4346-b6c0-47e3de576961): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:39:47 crc kubenswrapper[4930]: E0313 09:39:47.239213 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-6sjqb" podUID="94958e65-6d03-4346-b6c0-47e3de576961" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.592133 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.134:5671: i/o timeout" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.682555 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: E0313 09:39:47.688490 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-6sjqb" podUID="94958e65-6d03-4346-b6c0-47e3de576961" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.754197 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.786505 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.806233 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 09:39:47 crc kubenswrapper[4930]: E0313 09:39:47.807013 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" containerName="rabbitmq" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.807038 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" containerName="rabbitmq" Mar 13 09:39:47 crc kubenswrapper[4930]: E0313 09:39:47.807054 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" containerName="setup-container" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.807066 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" containerName="setup-container" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.807370 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" containerName="rabbitmq" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.809303 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.812787 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.813031 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-h595l" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.812980 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.813339 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.813560 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.813661 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.816461 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.826067 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.950693 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2392424a-e626-4305-865e-4817f46a9b65-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.950760 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2392424a-e626-4305-865e-4817f46a9b65-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.950856 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2392424a-e626-4305-865e-4817f46a9b65-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.950998 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24zp6\" (UniqueName: \"kubernetes.io/projected/2392424a-e626-4305-865e-4817f46a9b65-kube-api-access-24zp6\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.951103 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2392424a-e626-4305-865e-4817f46a9b65-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.951519 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2392424a-e626-4305-865e-4817f46a9b65-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.951616 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2392424a-e626-4305-865e-4817f46a9b65-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.951698 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2392424a-e626-4305-865e-4817f46a9b65-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.951796 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-cb39e555-44ae-4231-8aab-8695177f64b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.951940 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2392424a-e626-4305-865e-4817f46a9b65-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.952053 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2392424a-e626-4305-865e-4817f46a9b65-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.972304 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:39:47 crc kubenswrapper[4930]: E0313 09:39:47.972697 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:39:47 crc kubenswrapper[4930]: I0313 09:39:47.991916 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354" path="/var/lib/kubelet/pods/dc44fa4e-c7fd-4a4b-a03c-cb02e3b47354/volumes" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.054276 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2392424a-e626-4305-865e-4817f46a9b65-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.054402 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2392424a-e626-4305-865e-4817f46a9b65-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.054490 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2392424a-e626-4305-865e-4817f46a9b65-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.054553 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2392424a-e626-4305-865e-4817f46a9b65-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.054635 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-cb39e555-44ae-4231-8aab-8695177f64b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.054721 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2392424a-e626-4305-865e-4817f46a9b65-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.054802 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2392424a-e626-4305-865e-4817f46a9b65-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.054946 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2392424a-e626-4305-865e-4817f46a9b65-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.056520 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2392424a-e626-4305-865e-4817f46a9b65-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.056992 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2392424a-e626-4305-865e-4817f46a9b65-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.057025 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2392424a-e626-4305-865e-4817f46a9b65-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.057129 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2392424a-e626-4305-865e-4817f46a9b65-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.057161 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24zp6\" (UniqueName: \"kubernetes.io/projected/2392424a-e626-4305-865e-4817f46a9b65-kube-api-access-24zp6\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.057272 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2392424a-e626-4305-865e-4817f46a9b65-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.058235 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2392424a-e626-4305-865e-4817f46a9b65-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.058543 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2392424a-e626-4305-865e-4817f46a9b65-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.059034 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.059148 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-cb39e555-44ae-4231-8aab-8695177f64b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/31037259f8435172bc4f2806d5e4fc542b3e0cf5ddb9acb377b0574429b2c6a4/globalmount\"" pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.061614 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2392424a-e626-4305-865e-4817f46a9b65-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.062104 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2392424a-e626-4305-865e-4817f46a9b65-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.062607 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2392424a-e626-4305-865e-4817f46a9b65-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.079671 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2392424a-e626-4305-865e-4817f46a9b65-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.079805 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24zp6\" (UniqueName: \"kubernetes.io/projected/2392424a-e626-4305-865e-4817f46a9b65-kube-api-access-24zp6\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.151060 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-cb39e555-44ae-4231-8aab-8695177f64b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-cb39e555-44ae-4231-8aab-8695177f64b0\") pod \"rabbitmq-cell1-server-0\" (UID: \"2392424a-e626-4305-865e-4817f46a9b65\") " pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:48 crc kubenswrapper[4930]: I0313 09:39:48.161249 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:39:49 crc kubenswrapper[4930]: I0313 09:39:49.431946 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-n25jr"] Mar 13 09:39:49 crc kubenswrapper[4930]: E0313 09:39:49.566877 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Mar 13 09:39:49 crc kubenswrapper[4930]: E0313 09:39:49.566932 4930 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested" Mar 13 09:39:49 crc kubenswrapper[4930]: E0313 09:39:49.567040 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nbfh5bfhf4h598h599h9bh698h56h587hbfh8fh76h57ch565hdfhd4h597h577h97h54fh65fh5f6h655h588h58fh7bhcbh75h5b5h67fhb6h59q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-snkkd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(d4640978-a913-419b-9faa-8b230f5e51f2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 09:39:49 crc kubenswrapper[4930]: I0313 09:39:49.731885 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" event={"ID":"0a90a3cb-c199-4c04-a702-7cc29427a444","Type":"ContainerStarted","Data":"22433d188accf7dfcfae782349c03c26fe177bf2de5dfe94dae3be659ae3b15b"} Mar 13 09:39:50 crc kubenswrapper[4930]: I0313 09:39:50.075987 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-2"] Mar 13 09:39:50 crc kubenswrapper[4930]: W0313 09:39:50.079271 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc0f868a_2947_4e5a_a502_7ece4237551c.slice/crio-a94bb9296196dd7687e504c77c1ac8bd8b836e264d3351c6bff6a1e8aeeb26b4 WatchSource:0}: Error finding container a94bb9296196dd7687e504c77c1ac8bd8b836e264d3351c6bff6a1e8aeeb26b4: Status 404 returned error can't find the container with id a94bb9296196dd7687e504c77c1ac8bd8b836e264d3351c6bff6a1e8aeeb26b4 Mar 13 09:39:50 crc kubenswrapper[4930]: I0313 09:39:50.267944 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 13 09:39:50 crc kubenswrapper[4930]: I0313 09:39:50.750459 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"bc0f868a-2947-4e5a-a502-7ece4237551c","Type":"ContainerStarted","Data":"a94bb9296196dd7687e504c77c1ac8bd8b836e264d3351c6bff6a1e8aeeb26b4"} Mar 13 09:39:50 crc kubenswrapper[4930]: I0313 09:39:50.752175 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2392424a-e626-4305-865e-4817f46a9b65","Type":"ContainerStarted","Data":"e9188523e8af0bd90111ca3b6f8ec22b91ac186b8dc7f1173dbba691f6423be7"} Mar 13 09:39:50 crc kubenswrapper[4930]: I0313 09:39:50.754013 4930 generic.go:334] "Generic (PLEG): container finished" podID="0a90a3cb-c199-4c04-a702-7cc29427a444" containerID="e8b688e11c0b0529c34537a707689f9f547eed64a82faf6687f22d0758da8001" exitCode=0 Mar 13 09:39:50 crc kubenswrapper[4930]: I0313 09:39:50.754053 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" event={"ID":"0a90a3cb-c199-4c04-a702-7cc29427a444","Type":"ContainerDied","Data":"e8b688e11c0b0529c34537a707689f9f547eed64a82faf6687f22d0758da8001"} Mar 13 09:39:51 crc kubenswrapper[4930]: I0313 09:39:51.793689 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4640978-a913-419b-9faa-8b230f5e51f2","Type":"ContainerStarted","Data":"7ca789ad26da2a5f016d61206574d5af655b682285e87a2263695e67d2fbfa1a"} Mar 13 09:39:51 crc kubenswrapper[4930]: I0313 09:39:51.812747 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" event={"ID":"0a90a3cb-c199-4c04-a702-7cc29427a444","Type":"ContainerStarted","Data":"d13a84b42796b9b422aadf2ac30b56572c093c9e9a2921467fbae8a391f5b8b4"} Mar 13 09:39:51 crc kubenswrapper[4930]: I0313 09:39:51.815503 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:51 crc kubenswrapper[4930]: I0313 09:39:51.860100 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" podStartSLOduration=18.860073864 podStartE2EDuration="18.860073864s" podCreationTimestamp="2026-03-13 09:39:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:39:51.856585595 +0000 UTC m=+1632.606500272" watchObservedRunningTime="2026-03-13 09:39:51.860073864 +0000 UTC m=+1632.609988551" Mar 13 09:39:52 crc kubenswrapper[4930]: I0313 09:39:52.855158 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4640978-a913-419b-9faa-8b230f5e51f2","Type":"ContainerStarted","Data":"9c26120b518f3a76ed8b75222d28fee55d500545914ccd519b65a1dd5d5c3b14"} Mar 13 09:39:52 crc kubenswrapper[4930]: I0313 09:39:52.864256 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2392424a-e626-4305-865e-4817f46a9b65","Type":"ContainerStarted","Data":"ce5a2d9093e73f71c9d64e2de62aa7d1071e449498c9b7100c86258f471ed3cf"} Mar 13 09:39:52 crc kubenswrapper[4930]: I0313 09:39:52.873088 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"bc0f868a-2947-4e5a-a502-7ece4237551c","Type":"ContainerStarted","Data":"0530da87c60540412a729a5484f286cefc6122d24cacd7f82b54839defc2d427"} Mar 13 09:39:56 crc kubenswrapper[4930]: E0313 09:39:56.141907 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" Mar 13 09:39:56 crc kubenswrapper[4930]: I0313 09:39:56.920750 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4640978-a913-419b-9faa-8b230f5e51f2","Type":"ContainerStarted","Data":"3708622dae94ba1bab4db1b77a117f2a29cbd9c1db81d1c43a5ce2a5c8617e2d"} Mar 13 09:39:56 crc kubenswrapper[4930]: E0313 09:39:56.923648 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" Mar 13 09:39:56 crc kubenswrapper[4930]: I0313 09:39:56.924225 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 09:39:57 crc kubenswrapper[4930]: E0313 09:39:57.933646 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" Mar 13 09:39:58 crc kubenswrapper[4930]: E0313 09:39:58.947986 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-ceilometer-central:current-tested\\\"\"" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.064878 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.142392 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-5cbxc"] Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.142779 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" podUID="c11761f9-55b2-4645-ae32-e2a4caaa8bb6" containerName="dnsmasq-dns" containerID="cri-o://4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4" gracePeriod=10 Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.279427 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-27fl7"] Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.282399 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.298844 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-27fl7"] Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.388209 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.388303 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.388379 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.388511 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.388598 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-config\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.388764 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.388845 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb5sq\" (UniqueName: \"kubernetes.io/projected/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-kube-api-access-zb5sq\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.491810 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb5sq\" (UniqueName: \"kubernetes.io/projected/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-kube-api-access-zb5sq\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.492266 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.492335 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.492468 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.492530 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.492600 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-config\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.492634 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.493573 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-openstack-edpm-ipam\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.494150 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-ovsdbserver-sb\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.496484 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-dns-svc\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.496742 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-config\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.496817 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-dns-swift-storage-0\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.497791 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-ovsdbserver-nb\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.532860 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb5sq\" (UniqueName: \"kubernetes.io/projected/ffa41b0a-e7d4-4e37-ba32-3a52f73772c4-kube-api-access-zb5sq\") pod \"dnsmasq-dns-5d75f767dc-27fl7\" (UID: \"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4\") " pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.719199 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.916642 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:39:59 crc kubenswrapper[4930]: I0313 09:39:59.978916 4930 generic.go:334] "Generic (PLEG): container finished" podID="c11761f9-55b2-4645-ae32-e2a4caaa8bb6" containerID="4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4" exitCode=0 Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.043153 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9sxr2\" (UniqueName: \"kubernetes.io/projected/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-kube-api-access-9sxr2\") pod \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.046933 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-dns-svc\") pod \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.047082 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-ovsdbserver-sb\") pod \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.047292 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-dns-swift-storage-0\") pod \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.047419 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-config\") pod \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.047618 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-ovsdbserver-nb\") pod \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\" (UID: \"c11761f9-55b2-4645-ae32-e2a4caaa8bb6\") " Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.225171 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.241449 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" event={"ID":"c11761f9-55b2-4645-ae32-e2a4caaa8bb6","Type":"ContainerDied","Data":"4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4"} Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.241500 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-f84f9ccf-5cbxc" event={"ID":"c11761f9-55b2-4645-ae32-e2a4caaa8bb6","Type":"ContainerDied","Data":"5de233ac35bfb3ca3216fd1693bb31be51142d92255748e89ae9687bc6b36a6a"} Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.241532 4930 scope.go:117] "RemoveContainer" containerID="4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.260100 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-kube-api-access-9sxr2" (OuterVolumeSpecName: "kube-api-access-9sxr2") pod "c11761f9-55b2-4645-ae32-e2a4caaa8bb6" (UID: "c11761f9-55b2-4645-ae32-e2a4caaa8bb6"). InnerVolumeSpecName "kube-api-access-9sxr2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.311853 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9sxr2\" (UniqueName: \"kubernetes.io/projected/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-kube-api-access-9sxr2\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.316581 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c11761f9-55b2-4645-ae32-e2a4caaa8bb6" (UID: "c11761f9-55b2-4645-ae32-e2a4caaa8bb6"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.316826 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556580-xvwz8"] Mar 13 09:40:00 crc kubenswrapper[4930]: E0313 09:40:00.317488 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c11761f9-55b2-4645-ae32-e2a4caaa8bb6" containerName="dnsmasq-dns" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.317505 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c11761f9-55b2-4645-ae32-e2a4caaa8bb6" containerName="dnsmasq-dns" Mar 13 09:40:00 crc kubenswrapper[4930]: E0313 09:40:00.317538 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c11761f9-55b2-4645-ae32-e2a4caaa8bb6" containerName="init" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.317545 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c11761f9-55b2-4645-ae32-e2a4caaa8bb6" containerName="init" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.317858 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c11761f9-55b2-4645-ae32-e2a4caaa8bb6" containerName="dnsmasq-dns" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.318918 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556580-xvwz8" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.329022 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.333266 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.343100 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.369467 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c11761f9-55b2-4645-ae32-e2a4caaa8bb6" (UID: "c11761f9-55b2-4645-ae32-e2a4caaa8bb6"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.377504 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556580-xvwz8"] Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.414350 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.414392 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.421486 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c11761f9-55b2-4645-ae32-e2a4caaa8bb6" (UID: "c11761f9-55b2-4645-ae32-e2a4caaa8bb6"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.439910 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c11761f9-55b2-4645-ae32-e2a4caaa8bb6" (UID: "c11761f9-55b2-4645-ae32-e2a4caaa8bb6"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.443441 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-config" (OuterVolumeSpecName: "config") pod "c11761f9-55b2-4645-ae32-e2a4caaa8bb6" (UID: "c11761f9-55b2-4645-ae32-e2a4caaa8bb6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.516565 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkbps\" (UniqueName: \"kubernetes.io/projected/3d55c277-21b0-4e8f-a561-c20ba8c2ce8d-kube-api-access-hkbps\") pod \"auto-csr-approver-29556580-xvwz8\" (UID: \"3d55c277-21b0-4e8f-a561-c20ba8c2ce8d\") " pod="openshift-infra/auto-csr-approver-29556580-xvwz8" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.517523 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.517548 4930 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.517561 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c11761f9-55b2-4645-ae32-e2a4caaa8bb6-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.524603 4930 scope.go:117] "RemoveContainer" containerID="5d142d1645df31f75115da4a6211ab6b5bf9dacd0e7d33415b2c01edfbe54c2b" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.555800 4930 scope.go:117] "RemoveContainer" containerID="4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4" Mar 13 09:40:00 crc kubenswrapper[4930]: E0313 09:40:00.557163 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4\": container with ID starting with 4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4 not found: ID does not exist" containerID="4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.557337 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4"} err="failed to get container status \"4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4\": rpc error: code = NotFound desc = could not find container \"4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4\": container with ID starting with 4c94931cd8c54b7e92ee9ecf0b18350907ff3631ecd119a715daed11dd9546b4 not found: ID does not exist" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.557494 4930 scope.go:117] "RemoveContainer" containerID="5d142d1645df31f75115da4a6211ab6b5bf9dacd0e7d33415b2c01edfbe54c2b" Mar 13 09:40:00 crc kubenswrapper[4930]: E0313 09:40:00.558060 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d142d1645df31f75115da4a6211ab6b5bf9dacd0e7d33415b2c01edfbe54c2b\": container with ID starting with 5d142d1645df31f75115da4a6211ab6b5bf9dacd0e7d33415b2c01edfbe54c2b not found: ID does not exist" containerID="5d142d1645df31f75115da4a6211ab6b5bf9dacd0e7d33415b2c01edfbe54c2b" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.558083 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d142d1645df31f75115da4a6211ab6b5bf9dacd0e7d33415b2c01edfbe54c2b"} err="failed to get container status \"5d142d1645df31f75115da4a6211ab6b5bf9dacd0e7d33415b2c01edfbe54c2b\": rpc error: code = NotFound desc = could not find container \"5d142d1645df31f75115da4a6211ab6b5bf9dacd0e7d33415b2c01edfbe54c2b\": container with ID starting with 5d142d1645df31f75115da4a6211ab6b5bf9dacd0e7d33415b2c01edfbe54c2b not found: ID does not exist" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.619640 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkbps\" (UniqueName: \"kubernetes.io/projected/3d55c277-21b0-4e8f-a561-c20ba8c2ce8d-kube-api-access-hkbps\") pod \"auto-csr-approver-29556580-xvwz8\" (UID: \"3d55c277-21b0-4e8f-a561-c20ba8c2ce8d\") " pod="openshift-infra/auto-csr-approver-29556580-xvwz8" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.633135 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-5cbxc"] Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.638477 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkbps\" (UniqueName: \"kubernetes.io/projected/3d55c277-21b0-4e8f-a561-c20ba8c2ce8d-kube-api-access-hkbps\") pod \"auto-csr-approver-29556580-xvwz8\" (UID: \"3d55c277-21b0-4e8f-a561-c20ba8c2ce8d\") " pod="openshift-infra/auto-csr-approver-29556580-xvwz8" Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.645675 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-f84f9ccf-5cbxc"] Mar 13 09:40:00 crc kubenswrapper[4930]: W0313 09:40:00.670720 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podffa41b0a_e7d4_4e37_ba32_3a52f73772c4.slice/crio-865a2cf84e33a164c86de11113339ea85502b3f2ae4271ae4915d9459f20a5ce WatchSource:0}: Error finding container 865a2cf84e33a164c86de11113339ea85502b3f2ae4271ae4915d9459f20a5ce: Status 404 returned error can't find the container with id 865a2cf84e33a164c86de11113339ea85502b3f2ae4271ae4915d9459f20a5ce Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.685199 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5d75f767dc-27fl7"] Mar 13 09:40:00 crc kubenswrapper[4930]: I0313 09:40:00.833426 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556580-xvwz8" Mar 13 09:40:01 crc kubenswrapper[4930]: I0313 09:40:01.016135 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" event={"ID":"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4","Type":"ContainerStarted","Data":"865a2cf84e33a164c86de11113339ea85502b3f2ae4271ae4915d9459f20a5ce"} Mar 13 09:40:01 crc kubenswrapper[4930]: I0313 09:40:01.311841 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556580-xvwz8"] Mar 13 09:40:01 crc kubenswrapper[4930]: W0313 09:40:01.312620 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3d55c277_21b0_4e8f_a561_c20ba8c2ce8d.slice/crio-5ea75d07484fa77be2ceb6f9771adf9a9bd17e8d72624810215012e8729b5c5e WatchSource:0}: Error finding container 5ea75d07484fa77be2ceb6f9771adf9a9bd17e8d72624810215012e8729b5c5e: Status 404 returned error can't find the container with id 5ea75d07484fa77be2ceb6f9771adf9a9bd17e8d72624810215012e8729b5c5e Mar 13 09:40:01 crc kubenswrapper[4930]: I0313 09:40:01.972822 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:40:01 crc kubenswrapper[4930]: E0313 09:40:01.973210 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:40:01 crc kubenswrapper[4930]: I0313 09:40:01.988754 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c11761f9-55b2-4645-ae32-e2a4caaa8bb6" path="/var/lib/kubelet/pods/c11761f9-55b2-4645-ae32-e2a4caaa8bb6/volumes" Mar 13 09:40:02 crc kubenswrapper[4930]: I0313 09:40:02.029265 4930 generic.go:334] "Generic (PLEG): container finished" podID="ffa41b0a-e7d4-4e37-ba32-3a52f73772c4" containerID="5076ab5283cca6df8d80fca3ad94a53c5d28bd4674ac196813d7feb7c8dd412e" exitCode=0 Mar 13 09:40:02 crc kubenswrapper[4930]: I0313 09:40:02.029325 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" event={"ID":"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4","Type":"ContainerDied","Data":"5076ab5283cca6df8d80fca3ad94a53c5d28bd4674ac196813d7feb7c8dd412e"} Mar 13 09:40:02 crc kubenswrapper[4930]: I0313 09:40:02.030961 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556580-xvwz8" event={"ID":"3d55c277-21b0-4e8f-a561-c20ba8c2ce8d","Type":"ContainerStarted","Data":"5ea75d07484fa77be2ceb6f9771adf9a9bd17e8d72624810215012e8729b5c5e"} Mar 13 09:40:03 crc kubenswrapper[4930]: I0313 09:40:03.042831 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" event={"ID":"ffa41b0a-e7d4-4e37-ba32-3a52f73772c4","Type":"ContainerStarted","Data":"f12cafc6ea5e3c595c894c0681bc4070245dae9e5951d566555874dadd2cc46b"} Mar 13 09:40:03 crc kubenswrapper[4930]: I0313 09:40:03.043264 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:40:03 crc kubenswrapper[4930]: I0313 09:40:03.066450 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" podStartSLOduration=4.066419786 podStartE2EDuration="4.066419786s" podCreationTimestamp="2026-03-13 09:39:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:40:03.06223909 +0000 UTC m=+1643.812153777" watchObservedRunningTime="2026-03-13 09:40:03.066419786 +0000 UTC m=+1643.816334463" Mar 13 09:40:08 crc kubenswrapper[4930]: I0313 09:40:08.095536 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6sjqb" event={"ID":"94958e65-6d03-4346-b6c0-47e3de576961","Type":"ContainerStarted","Data":"73fd01a431aee4e9e432af3440c1d2dd1c2c77f8d2ac43502cac74a3c95ceff6"} Mar 13 09:40:09 crc kubenswrapper[4930]: I0313 09:40:09.109860 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556580-xvwz8" event={"ID":"3d55c277-21b0-4e8f-a561-c20ba8c2ce8d","Type":"ContainerStarted","Data":"3c1f9e12a3d90e441e6432e6727e9aa72a5dac133603645f94a85129d8be0a90"} Mar 13 09:40:09 crc kubenswrapper[4930]: I0313 09:40:09.134743 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556580-xvwz8" podStartSLOduration=3.215619635 podStartE2EDuration="9.134723871s" podCreationTimestamp="2026-03-13 09:40:00 +0000 UTC" firstStartedPulling="2026-03-13 09:40:01.315795381 +0000 UTC m=+1642.065710058" lastFinishedPulling="2026-03-13 09:40:07.234899617 +0000 UTC m=+1647.984814294" observedRunningTime="2026-03-13 09:40:09.128160615 +0000 UTC m=+1649.878075292" watchObservedRunningTime="2026-03-13 09:40:09.134723871 +0000 UTC m=+1649.884638548" Mar 13 09:40:09 crc kubenswrapper[4930]: I0313 09:40:09.158814 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-6sjqb" podStartSLOduration=3.861887045 podStartE2EDuration="56.158792792s" podCreationTimestamp="2026-03-13 09:39:13 +0000 UTC" firstStartedPulling="2026-03-13 09:39:14.937182569 +0000 UTC m=+1595.687097246" lastFinishedPulling="2026-03-13 09:40:07.234088296 +0000 UTC m=+1647.984002993" observedRunningTime="2026-03-13 09:40:09.148289456 +0000 UTC m=+1649.898204133" watchObservedRunningTime="2026-03-13 09:40:09.158792792 +0000 UTC m=+1649.908707459" Mar 13 09:40:09 crc kubenswrapper[4930]: I0313 09:40:09.722210 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5d75f767dc-27fl7" Mar 13 09:40:09 crc kubenswrapper[4930]: I0313 09:40:09.814062 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-n25jr"] Mar 13 09:40:09 crc kubenswrapper[4930]: I0313 09:40:09.814814 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" podUID="0a90a3cb-c199-4c04-a702-7cc29427a444" containerName="dnsmasq-dns" containerID="cri-o://d13a84b42796b9b422aadf2ac30b56572c093c9e9a2921467fbae8a391f5b8b4" gracePeriod=10 Mar 13 09:40:10 crc kubenswrapper[4930]: I0313 09:40:10.010699 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 09:40:10 crc kubenswrapper[4930]: I0313 09:40:10.124150 4930 generic.go:334] "Generic (PLEG): container finished" podID="0a90a3cb-c199-4c04-a702-7cc29427a444" containerID="d13a84b42796b9b422aadf2ac30b56572c093c9e9a2921467fbae8a391f5b8b4" exitCode=0 Mar 13 09:40:10 crc kubenswrapper[4930]: I0313 09:40:10.124228 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" event={"ID":"0a90a3cb-c199-4c04-a702-7cc29427a444","Type":"ContainerDied","Data":"d13a84b42796b9b422aadf2ac30b56572c093c9e9a2921467fbae8a391f5b8b4"} Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.615454 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.730863 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-ovsdbserver-nb\") pod \"0a90a3cb-c199-4c04-a702-7cc29427a444\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.731324 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-dns-swift-storage-0\") pod \"0a90a3cb-c199-4c04-a702-7cc29427a444\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.731489 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-openstack-edpm-ipam\") pod \"0a90a3cb-c199-4c04-a702-7cc29427a444\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.731540 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84dtz\" (UniqueName: \"kubernetes.io/projected/0a90a3cb-c199-4c04-a702-7cc29427a444-kube-api-access-84dtz\") pod \"0a90a3cb-c199-4c04-a702-7cc29427a444\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.731654 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-ovsdbserver-sb\") pod \"0a90a3cb-c199-4c04-a702-7cc29427a444\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.731700 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-config\") pod \"0a90a3cb-c199-4c04-a702-7cc29427a444\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.731742 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-dns-svc\") pod \"0a90a3cb-c199-4c04-a702-7cc29427a444\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.748052 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a90a3cb-c199-4c04-a702-7cc29427a444-kube-api-access-84dtz" (OuterVolumeSpecName: "kube-api-access-84dtz") pod "0a90a3cb-c199-4c04-a702-7cc29427a444" (UID: "0a90a3cb-c199-4c04-a702-7cc29427a444"). InnerVolumeSpecName "kube-api-access-84dtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.796850 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0a90a3cb-c199-4c04-a702-7cc29427a444" (UID: "0a90a3cb-c199-4c04-a702-7cc29427a444"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.799702 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0a90a3cb-c199-4c04-a702-7cc29427a444" (UID: "0a90a3cb-c199-4c04-a702-7cc29427a444"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.801115 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0a90a3cb-c199-4c04-a702-7cc29427a444" (UID: "0a90a3cb-c199-4c04-a702-7cc29427a444"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.802090 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "0a90a3cb-c199-4c04-a702-7cc29427a444" (UID: "0a90a3cb-c199-4c04-a702-7cc29427a444"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:12 crc kubenswrapper[4930]: E0313 09:40:12.805044 4930 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-config podName:0a90a3cb-c199-4c04-a702-7cc29427a444 nodeName:}" failed. No retries permitted until 2026-03-13 09:40:13.305014733 +0000 UTC m=+1654.054929490 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-config") pod "0a90a3cb-c199-4c04-a702-7cc29427a444" (UID: "0a90a3cb-c199-4c04-a702-7cc29427a444") : error deleting /var/lib/kubelet/pods/0a90a3cb-c199-4c04-a702-7cc29427a444/volume-subpaths: remove /var/lib/kubelet/pods/0a90a3cb-c199-4c04-a702-7cc29427a444/volume-subpaths: no such file or directory Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.805376 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0a90a3cb-c199-4c04-a702-7cc29427a444" (UID: "0a90a3cb-c199-4c04-a702-7cc29427a444"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.834861 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.834893 4930 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.834902 4930 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.834911 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84dtz\" (UniqueName: \"kubernetes.io/projected/0a90a3cb-c199-4c04-a702-7cc29427a444-kube-api-access-84dtz\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.834922 4930 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:12 crc kubenswrapper[4930]: I0313 09:40:12.834931 4930 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:13 crc kubenswrapper[4930]: I0313 09:40:13.178349 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" event={"ID":"0a90a3cb-c199-4c04-a702-7cc29427a444","Type":"ContainerDied","Data":"22433d188accf7dfcfae782349c03c26fe177bf2de5dfe94dae3be659ae3b15b"} Mar 13 09:40:13 crc kubenswrapper[4930]: I0313 09:40:13.178402 4930 scope.go:117] "RemoveContainer" containerID="d13a84b42796b9b422aadf2ac30b56572c093c9e9a2921467fbae8a391f5b8b4" Mar 13 09:40:13 crc kubenswrapper[4930]: I0313 09:40:13.178411 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b75489c6f-n25jr" Mar 13 09:40:13 crc kubenswrapper[4930]: I0313 09:40:13.214282 4930 scope.go:117] "RemoveContainer" containerID="e8b688e11c0b0529c34537a707689f9f547eed64a82faf6687f22d0758da8001" Mar 13 09:40:13 crc kubenswrapper[4930]: I0313 09:40:13.346285 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-config\") pod \"0a90a3cb-c199-4c04-a702-7cc29427a444\" (UID: \"0a90a3cb-c199-4c04-a702-7cc29427a444\") " Mar 13 09:40:13 crc kubenswrapper[4930]: I0313 09:40:13.346834 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-config" (OuterVolumeSpecName: "config") pod "0a90a3cb-c199-4c04-a702-7cc29427a444" (UID: "0a90a3cb-c199-4c04-a702-7cc29427a444"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:13 crc kubenswrapper[4930]: I0313 09:40:13.347362 4930 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0a90a3cb-c199-4c04-a702-7cc29427a444-config\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:13 crc kubenswrapper[4930]: I0313 09:40:13.522893 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-n25jr"] Mar 13 09:40:13 crc kubenswrapper[4930]: I0313 09:40:13.534355 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b75489c6f-n25jr"] Mar 13 09:40:14 crc kubenswrapper[4930]: I0313 09:40:14.000964 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a90a3cb-c199-4c04-a702-7cc29427a444" path="/var/lib/kubelet/pods/0a90a3cb-c199-4c04-a702-7cc29427a444/volumes" Mar 13 09:40:16 crc kubenswrapper[4930]: I0313 09:40:16.217192 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4640978-a913-419b-9faa-8b230f5e51f2","Type":"ContainerStarted","Data":"1cd5e3669981d79173656007193a66857ed29e6d11519fa9bc1873410a276c0f"} Mar 13 09:40:16 crc kubenswrapper[4930]: I0313 09:40:16.251653 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.413212523 podStartE2EDuration="55.251624419s" podCreationTimestamp="2026-03-13 09:39:21 +0000 UTC" firstStartedPulling="2026-03-13 09:39:22.557191239 +0000 UTC m=+1603.307105926" lastFinishedPulling="2026-03-13 09:40:15.395603145 +0000 UTC m=+1656.145517822" observedRunningTime="2026-03-13 09:40:16.241856901 +0000 UTC m=+1656.991771588" watchObservedRunningTime="2026-03-13 09:40:16.251624419 +0000 UTC m=+1657.001539096" Mar 13 09:40:16 crc kubenswrapper[4930]: I0313 09:40:16.971917 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:40:16 crc kubenswrapper[4930]: E0313 09:40:16.972416 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:40:17 crc kubenswrapper[4930]: I0313 09:40:17.247026 4930 generic.go:334] "Generic (PLEG): container finished" podID="3d55c277-21b0-4e8f-a561-c20ba8c2ce8d" containerID="3c1f9e12a3d90e441e6432e6727e9aa72a5dac133603645f94a85129d8be0a90" exitCode=0 Mar 13 09:40:17 crc kubenswrapper[4930]: I0313 09:40:17.247118 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556580-xvwz8" event={"ID":"3d55c277-21b0-4e8f-a561-c20ba8c2ce8d","Type":"ContainerDied","Data":"3c1f9e12a3d90e441e6432e6727e9aa72a5dac133603645f94a85129d8be0a90"} Mar 13 09:40:17 crc kubenswrapper[4930]: I0313 09:40:17.251547 4930 generic.go:334] "Generic (PLEG): container finished" podID="94958e65-6d03-4346-b6c0-47e3de576961" containerID="73fd01a431aee4e9e432af3440c1d2dd1c2c77f8d2ac43502cac74a3c95ceff6" exitCode=0 Mar 13 09:40:17 crc kubenswrapper[4930]: I0313 09:40:17.251591 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6sjqb" event={"ID":"94958e65-6d03-4346-b6c0-47e3de576961","Type":"ContainerDied","Data":"73fd01a431aee4e9e432af3440c1d2dd1c2c77f8d2ac43502cac74a3c95ceff6"} Mar 13 09:40:18 crc kubenswrapper[4930]: I0313 09:40:18.842405 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556580-xvwz8" Mar 13 09:40:18 crc kubenswrapper[4930]: I0313 09:40:18.910277 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hkbps\" (UniqueName: \"kubernetes.io/projected/3d55c277-21b0-4e8f-a561-c20ba8c2ce8d-kube-api-access-hkbps\") pod \"3d55c277-21b0-4e8f-a561-c20ba8c2ce8d\" (UID: \"3d55c277-21b0-4e8f-a561-c20ba8c2ce8d\") " Mar 13 09:40:18 crc kubenswrapper[4930]: I0313 09:40:18.937195 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d55c277-21b0-4e8f-a561-c20ba8c2ce8d-kube-api-access-hkbps" (OuterVolumeSpecName: "kube-api-access-hkbps") pod "3d55c277-21b0-4e8f-a561-c20ba8c2ce8d" (UID: "3d55c277-21b0-4e8f-a561-c20ba8c2ce8d"). InnerVolumeSpecName "kube-api-access-hkbps". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.023247 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hkbps\" (UniqueName: \"kubernetes.io/projected/3d55c277-21b0-4e8f-a561-c20ba8c2ce8d-kube-api-access-hkbps\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.132806 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6sjqb" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.227173 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94958e65-6d03-4346-b6c0-47e3de576961-config-data\") pod \"94958e65-6d03-4346-b6c0-47e3de576961\" (UID: \"94958e65-6d03-4346-b6c0-47e3de576961\") " Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.227654 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94958e65-6d03-4346-b6c0-47e3de576961-combined-ca-bundle\") pod \"94958e65-6d03-4346-b6c0-47e3de576961\" (UID: \"94958e65-6d03-4346-b6c0-47e3de576961\") " Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.227752 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gm7g\" (UniqueName: \"kubernetes.io/projected/94958e65-6d03-4346-b6c0-47e3de576961-kube-api-access-7gm7g\") pod \"94958e65-6d03-4346-b6c0-47e3de576961\" (UID: \"94958e65-6d03-4346-b6c0-47e3de576961\") " Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.233198 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94958e65-6d03-4346-b6c0-47e3de576961-kube-api-access-7gm7g" (OuterVolumeSpecName: "kube-api-access-7gm7g") pod "94958e65-6d03-4346-b6c0-47e3de576961" (UID: "94958e65-6d03-4346-b6c0-47e3de576961"). InnerVolumeSpecName "kube-api-access-7gm7g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.268150 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94958e65-6d03-4346-b6c0-47e3de576961-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "94958e65-6d03-4346-b6c0-47e3de576961" (UID: "94958e65-6d03-4346-b6c0-47e3de576961"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.288825 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556580-xvwz8" event={"ID":"3d55c277-21b0-4e8f-a561-c20ba8c2ce8d","Type":"ContainerDied","Data":"5ea75d07484fa77be2ceb6f9771adf9a9bd17e8d72624810215012e8729b5c5e"} Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.288866 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5ea75d07484fa77be2ceb6f9771adf9a9bd17e8d72624810215012e8729b5c5e" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.288941 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556580-xvwz8" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.306166 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6sjqb" event={"ID":"94958e65-6d03-4346-b6c0-47e3de576961","Type":"ContainerDied","Data":"b713f4a7271e88352615848e0355fa77409615c8dc2f9f940c3702e1ee0fc1bb"} Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.306209 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b713f4a7271e88352615848e0355fa77409615c8dc2f9f940c3702e1ee0fc1bb" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.306248 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6sjqb" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.335169 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/94958e65-6d03-4346-b6c0-47e3de576961-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.335499 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gm7g\" (UniqueName: \"kubernetes.io/projected/94958e65-6d03-4346-b6c0-47e3de576961-kube-api-access-7gm7g\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.344181 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94958e65-6d03-4346-b6c0-47e3de576961-config-data" (OuterVolumeSpecName: "config-data") pod "94958e65-6d03-4346-b6c0-47e3de576961" (UID: "94958e65-6d03-4346-b6c0-47e3de576961"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.361868 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556574-dj8l8"] Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.380782 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556574-dj8l8"] Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.438203 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/94958e65-6d03-4346-b6c0-47e3de576961-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:19 crc kubenswrapper[4930]: I0313 09:40:19.984963 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2b7d7fd-2404-409b-a6bb-c000f04536cb" path="/var/lib/kubelet/pods/a2b7d7fd-2404-409b-a6bb-c000f04536cb/volumes" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.079148 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx"] Mar 13 09:40:20 crc kubenswrapper[4930]: E0313 09:40:20.080156 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a90a3cb-c199-4c04-a702-7cc29427a444" containerName="dnsmasq-dns" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.080169 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a90a3cb-c199-4c04-a702-7cc29427a444" containerName="dnsmasq-dns" Mar 13 09:40:20 crc kubenswrapper[4930]: E0313 09:40:20.080199 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d55c277-21b0-4e8f-a561-c20ba8c2ce8d" containerName="oc" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.080210 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d55c277-21b0-4e8f-a561-c20ba8c2ce8d" containerName="oc" Mar 13 09:40:20 crc kubenswrapper[4930]: E0313 09:40:20.080230 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a90a3cb-c199-4c04-a702-7cc29427a444" containerName="init" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.080238 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a90a3cb-c199-4c04-a702-7cc29427a444" containerName="init" Mar 13 09:40:20 crc kubenswrapper[4930]: E0313 09:40:20.080255 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94958e65-6d03-4346-b6c0-47e3de576961" containerName="heat-db-sync" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.080263 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="94958e65-6d03-4346-b6c0-47e3de576961" containerName="heat-db-sync" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.080483 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a90a3cb-c199-4c04-a702-7cc29427a444" containerName="dnsmasq-dns" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.080505 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="94958e65-6d03-4346-b6c0-47e3de576961" containerName="heat-db-sync" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.080521 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d55c277-21b0-4e8f-a561-c20ba8c2ce8d" containerName="oc" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.081323 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.085115 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.085170 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.085413 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.085609 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.122829 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx"] Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.157506 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.157603 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.157793 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4lqv\" (UniqueName: \"kubernetes.io/projected/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-kube-api-access-w4lqv\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.157814 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.260510 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4lqv\" (UniqueName: \"kubernetes.io/projected/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-kube-api-access-w4lqv\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.260564 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.260668 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.260727 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.268658 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.269091 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.272182 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.278334 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4lqv\" (UniqueName: \"kubernetes.io/projected/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-kube-api-access-w4lqv\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.401719 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.729555 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-f9949f6d4-jldnd"] Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.732092 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.755713 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-f9949f6d4-jldnd"] Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.776807 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8e68bf4-ac94-446c-883a-758ad1dfdb89-config-data\") pod \"heat-engine-f9949f6d4-jldnd\" (UID: \"c8e68bf4-ac94-446c-883a-758ad1dfdb89\") " pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.776914 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8e68bf4-ac94-446c-883a-758ad1dfdb89-combined-ca-bundle\") pod \"heat-engine-f9949f6d4-jldnd\" (UID: \"c8e68bf4-ac94-446c-883a-758ad1dfdb89\") " pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.777027 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-89ghc\" (UniqueName: \"kubernetes.io/projected/c8e68bf4-ac94-446c-883a-758ad1dfdb89-kube-api-access-89ghc\") pod \"heat-engine-f9949f6d4-jldnd\" (UID: \"c8e68bf4-ac94-446c-883a-758ad1dfdb89\") " pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.777092 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8e68bf4-ac94-446c-883a-758ad1dfdb89-config-data-custom\") pod \"heat-engine-f9949f6d4-jldnd\" (UID: \"c8e68bf4-ac94-446c-883a-758ad1dfdb89\") " pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.783036 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-575b8b7644-p4hwx"] Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.784984 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.812703 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-575b8b7644-p4hwx"] Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.826170 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-6c996bbf9d-k5gp9"] Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.828097 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.854170 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6c996bbf9d-k5gp9"] Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.878734 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-public-tls-certs\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.878781 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-89ghc\" (UniqueName: \"kubernetes.io/projected/c8e68bf4-ac94-446c-883a-758ad1dfdb89-kube-api-access-89ghc\") pod \"heat-engine-f9949f6d4-jldnd\" (UID: \"c8e68bf4-ac94-446c-883a-758ad1dfdb89\") " pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.878837 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8e68bf4-ac94-446c-883a-758ad1dfdb89-config-data-custom\") pod \"heat-engine-f9949f6d4-jldnd\" (UID: \"c8e68bf4-ac94-446c-883a-758ad1dfdb89\") " pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.878856 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-config-data\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.878873 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-config-data-custom\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.878894 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-public-tls-certs\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.878925 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-config-data\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.878948 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-internal-tls-certs\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.878978 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-internal-tls-certs\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.879001 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8e68bf4-ac94-446c-883a-758ad1dfdb89-config-data\") pod \"heat-engine-f9949f6d4-jldnd\" (UID: \"c8e68bf4-ac94-446c-883a-758ad1dfdb89\") " pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.879054 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cr2nx\" (UniqueName: \"kubernetes.io/projected/4a79e474-ae50-4d4e-a809-29033d1357e9-kube-api-access-cr2nx\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.879097 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vg97\" (UniqueName: \"kubernetes.io/projected/95d23f86-d860-42ac-9b8d-edf7854096d2-kube-api-access-2vg97\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.879130 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-combined-ca-bundle\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.879154 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8e68bf4-ac94-446c-883a-758ad1dfdb89-combined-ca-bundle\") pod \"heat-engine-f9949f6d4-jldnd\" (UID: \"c8e68bf4-ac94-446c-883a-758ad1dfdb89\") " pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.879189 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-config-data-custom\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.879235 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-combined-ca-bundle\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.889495 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c8e68bf4-ac94-446c-883a-758ad1dfdb89-config-data-custom\") pod \"heat-engine-f9949f6d4-jldnd\" (UID: \"c8e68bf4-ac94-446c-883a-758ad1dfdb89\") " pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.892164 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8e68bf4-ac94-446c-883a-758ad1dfdb89-config-data\") pod \"heat-engine-f9949f6d4-jldnd\" (UID: \"c8e68bf4-ac94-446c-883a-758ad1dfdb89\") " pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.893041 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8e68bf4-ac94-446c-883a-758ad1dfdb89-combined-ca-bundle\") pod \"heat-engine-f9949f6d4-jldnd\" (UID: \"c8e68bf4-ac94-446c-883a-758ad1dfdb89\") " pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.896985 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-89ghc\" (UniqueName: \"kubernetes.io/projected/c8e68bf4-ac94-446c-883a-758ad1dfdb89-kube-api-access-89ghc\") pod \"heat-engine-f9949f6d4-jldnd\" (UID: \"c8e68bf4-ac94-446c-883a-758ad1dfdb89\") " pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.980929 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-config-data\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.980984 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-internal-tls-certs\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.981024 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-internal-tls-certs\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.981094 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cr2nx\" (UniqueName: \"kubernetes.io/projected/4a79e474-ae50-4d4e-a809-29033d1357e9-kube-api-access-cr2nx\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.981133 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vg97\" (UniqueName: \"kubernetes.io/projected/95d23f86-d860-42ac-9b8d-edf7854096d2-kube-api-access-2vg97\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.981166 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-combined-ca-bundle\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.981210 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-config-data-custom\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.981227 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-combined-ca-bundle\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.981305 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-public-tls-certs\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.981374 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-config-data\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.981392 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-config-data-custom\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.981409 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-public-tls-certs\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.987042 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-public-tls-certs\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.987320 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-combined-ca-bundle\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.988878 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-config-data\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.991930 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-public-tls-certs\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:20 crc kubenswrapper[4930]: I0313 09:40:20.993847 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-internal-tls-certs\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:21 crc kubenswrapper[4930]: I0313 09:40:21.014710 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/95d23f86-d860-42ac-9b8d-edf7854096d2-config-data-custom\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:21 crc kubenswrapper[4930]: I0313 09:40:21.014793 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-combined-ca-bundle\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:21 crc kubenswrapper[4930]: I0313 09:40:21.014898 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-internal-tls-certs\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:21 crc kubenswrapper[4930]: I0313 09:40:21.016260 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-config-data\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:21 crc kubenswrapper[4930]: I0313 09:40:21.016943 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4a79e474-ae50-4d4e-a809-29033d1357e9-config-data-custom\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:21 crc kubenswrapper[4930]: I0313 09:40:21.019198 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cr2nx\" (UniqueName: \"kubernetes.io/projected/4a79e474-ae50-4d4e-a809-29033d1357e9-kube-api-access-cr2nx\") pod \"heat-cfnapi-6c996bbf9d-k5gp9\" (UID: \"4a79e474-ae50-4d4e-a809-29033d1357e9\") " pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:21 crc kubenswrapper[4930]: I0313 09:40:21.019814 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vg97\" (UniqueName: \"kubernetes.io/projected/95d23f86-d860-42ac-9b8d-edf7854096d2-kube-api-access-2vg97\") pod \"heat-api-575b8b7644-p4hwx\" (UID: \"95d23f86-d860-42ac-9b8d-edf7854096d2\") " pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:21 crc kubenswrapper[4930]: I0313 09:40:21.068912 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:21 crc kubenswrapper[4930]: I0313 09:40:21.114414 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:21 crc kubenswrapper[4930]: I0313 09:40:21.158161 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:21 crc kubenswrapper[4930]: I0313 09:40:21.709885 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-f9949f6d4-jldnd"] Mar 13 09:40:21 crc kubenswrapper[4930]: I0313 09:40:21.911121 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx"] Mar 13 09:40:21 crc kubenswrapper[4930]: W0313 09:40:21.964061 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode68e5869_c983_4ae1_a8c7_d6c4fba11d30.slice/crio-d794449ec86f8a8907058a764570206d3641bab10a4709aa32632690394d7c67 WatchSource:0}: Error finding container d794449ec86f8a8907058a764570206d3641bab10a4709aa32632690394d7c67: Status 404 returned error can't find the container with id d794449ec86f8a8907058a764570206d3641bab10a4709aa32632690394d7c67 Mar 13 09:40:22 crc kubenswrapper[4930]: I0313 09:40:21.978154 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 09:40:22 crc kubenswrapper[4930]: I0313 09:40:22.029882 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-6c996bbf9d-k5gp9"] Mar 13 09:40:22 crc kubenswrapper[4930]: I0313 09:40:22.029936 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-575b8b7644-p4hwx"] Mar 13 09:40:22 crc kubenswrapper[4930]: I0313 09:40:22.362700 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" event={"ID":"e68e5869-c983-4ae1-a8c7-d6c4fba11d30","Type":"ContainerStarted","Data":"d794449ec86f8a8907058a764570206d3641bab10a4709aa32632690394d7c67"} Mar 13 09:40:22 crc kubenswrapper[4930]: I0313 09:40:22.365848 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-f9949f6d4-jldnd" event={"ID":"c8e68bf4-ac94-446c-883a-758ad1dfdb89","Type":"ContainerStarted","Data":"7100724ba0ccc3535203842d813dda9a6a3f8033845620fe8392a50bdbae9c2a"} Mar 13 09:40:22 crc kubenswrapper[4930]: I0313 09:40:22.365876 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-f9949f6d4-jldnd" event={"ID":"c8e68bf4-ac94-446c-883a-758ad1dfdb89","Type":"ContainerStarted","Data":"99ff731475dca1c3d9533f530be06050d688e8e7b044598da84a34edc48a02c0"} Mar 13 09:40:22 crc kubenswrapper[4930]: I0313 09:40:22.366018 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:22 crc kubenswrapper[4930]: I0313 09:40:22.369552 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" event={"ID":"4a79e474-ae50-4d4e-a809-29033d1357e9","Type":"ContainerStarted","Data":"e25b583f7142dbeb670980327080583b7d548c240cef654e1c8f2c42d3d3f084"} Mar 13 09:40:22 crc kubenswrapper[4930]: I0313 09:40:22.372125 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-575b8b7644-p4hwx" event={"ID":"95d23f86-d860-42ac-9b8d-edf7854096d2","Type":"ContainerStarted","Data":"757d7a06900c3bb259cc3792298f09f7cdb4d816abb76c2aaa20c8b231b45e93"} Mar 13 09:40:22 crc kubenswrapper[4930]: I0313 09:40:22.394058 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-f9949f6d4-jldnd" podStartSLOduration=2.394034994 podStartE2EDuration="2.394034994s" podCreationTimestamp="2026-03-13 09:40:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:40:22.386970505 +0000 UTC m=+1663.136885192" watchObservedRunningTime="2026-03-13 09:40:22.394034994 +0000 UTC m=+1663.143949671" Mar 13 09:40:24 crc kubenswrapper[4930]: I0313 09:40:24.413164 4930 generic.go:334] "Generic (PLEG): container finished" podID="bc0f868a-2947-4e5a-a502-7ece4237551c" containerID="0530da87c60540412a729a5484f286cefc6122d24cacd7f82b54839defc2d427" exitCode=0 Mar 13 09:40:24 crc kubenswrapper[4930]: I0313 09:40:24.413302 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"bc0f868a-2947-4e5a-a502-7ece4237551c","Type":"ContainerDied","Data":"0530da87c60540412a729a5484f286cefc6122d24cacd7f82b54839defc2d427"} Mar 13 09:40:24 crc kubenswrapper[4930]: I0313 09:40:24.585108 4930 scope.go:117] "RemoveContainer" containerID="90290252b81588f175401788aba185e7d3a83628e294a972697ba4b38ed16e21" Mar 13 09:40:25 crc kubenswrapper[4930]: I0313 09:40:25.427798 4930 generic.go:334] "Generic (PLEG): container finished" podID="2392424a-e626-4305-865e-4817f46a9b65" containerID="ce5a2d9093e73f71c9d64e2de62aa7d1071e449498c9b7100c86258f471ed3cf" exitCode=0 Mar 13 09:40:25 crc kubenswrapper[4930]: I0313 09:40:25.427917 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2392424a-e626-4305-865e-4817f46a9b65","Type":"ContainerDied","Data":"ce5a2d9093e73f71c9d64e2de62aa7d1071e449498c9b7100c86258f471ed3cf"} Mar 13 09:40:26 crc kubenswrapper[4930]: I0313 09:40:26.082824 4930 scope.go:117] "RemoveContainer" containerID="12f5210598b95db7dbe92711d04a787ad629536cd28bb916dd8442d27c127c93" Mar 13 09:40:26 crc kubenswrapper[4930]: I0313 09:40:26.361127 4930 scope.go:117] "RemoveContainer" containerID="544f3748e5dc2a3c11f5bae45b8a9e44d286062863d4c6013cca7699288a5348" Mar 13 09:40:26 crc kubenswrapper[4930]: I0313 09:40:26.441050 4930 scope.go:117] "RemoveContainer" containerID="0ff24caf2459e3a24991ce35ffc0618b23e74e2db943bd600c2e1db6808a42da" Mar 13 09:40:26 crc kubenswrapper[4930]: I0313 09:40:26.501746 4930 scope.go:117] "RemoveContainer" containerID="5d93f52f3a7b4038546d2005345570ab31cd4165cbad2ec0b14428854e754b01" Mar 13 09:40:26 crc kubenswrapper[4930]: I0313 09:40:26.536062 4930 scope.go:117] "RemoveContainer" containerID="43fb9ce2d8aa5390976a443491d5cf85b350d3f0f72289a47f81a4a7564022d1" Mar 13 09:40:26 crc kubenswrapper[4930]: I0313 09:40:26.613114 4930 scope.go:117] "RemoveContainer" containerID="be9f2773447a84d88c2d62d4824149378dce13cfe4f15b91538bba9477a016e6" Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.463125 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"2392424a-e626-4305-865e-4817f46a9b65","Type":"ContainerStarted","Data":"6805269df30139a79edfd06d0d968559553767481f026b5f271ec6aa75fc763e"} Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.465072 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.468711 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" event={"ID":"4a79e474-ae50-4d4e-a809-29033d1357e9","Type":"ContainerStarted","Data":"8a912db57a55a2ddc96391d3f735c626c3195c110ed672bec9dd5e1493aab567"} Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.468832 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.470559 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-2" event={"ID":"bc0f868a-2947-4e5a-a502-7ece4237551c","Type":"ContainerStarted","Data":"6ab5fb0e4559abfb3a724426d4b732260e2ae249ab49d079140411dba3ff633b"} Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.470832 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-2" Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.473065 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-575b8b7644-p4hwx" event={"ID":"95d23f86-d860-42ac-9b8d-edf7854096d2","Type":"ContainerStarted","Data":"9016cfa532da4ee37f909ecbe8893f724d0a7a3f2c22377da78d2ee37df03490"} Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.473598 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.496889 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.496869577 podStartE2EDuration="40.496869577s" podCreationTimestamp="2026-03-13 09:39:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:40:27.492608439 +0000 UTC m=+1668.242523116" watchObservedRunningTime="2026-03-13 09:40:27.496869577 +0000 UTC m=+1668.246784254" Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.534498 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-2" podStartSLOduration=52.534473612 podStartE2EDuration="52.534473612s" podCreationTimestamp="2026-03-13 09:39:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:40:27.524273153 +0000 UTC m=+1668.274187830" watchObservedRunningTime="2026-03-13 09:40:27.534473612 +0000 UTC m=+1668.284388309" Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.570699 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" podStartSLOduration=3.516996943 podStartE2EDuration="7.5706572s" podCreationTimestamp="2026-03-13 09:40:20 +0000 UTC" firstStartedPulling="2026-03-13 09:40:22.028994886 +0000 UTC m=+1662.778909563" lastFinishedPulling="2026-03-13 09:40:26.082655133 +0000 UTC m=+1666.832569820" observedRunningTime="2026-03-13 09:40:27.548284952 +0000 UTC m=+1668.298199629" watchObservedRunningTime="2026-03-13 09:40:27.5706572 +0000 UTC m=+1668.320571877" Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.603393 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-575b8b7644-p4hwx" podStartSLOduration=3.568900322 podStartE2EDuration="7.603368311s" podCreationTimestamp="2026-03-13 09:40:20 +0000 UTC" firstStartedPulling="2026-03-13 09:40:22.048569173 +0000 UTC m=+1662.798483850" lastFinishedPulling="2026-03-13 09:40:26.083037172 +0000 UTC m=+1666.832951839" observedRunningTime="2026-03-13 09:40:27.571022919 +0000 UTC m=+1668.320937596" watchObservedRunningTime="2026-03-13 09:40:27.603368311 +0000 UTC m=+1668.353282988" Mar 13 09:40:27 crc kubenswrapper[4930]: I0313 09:40:27.970848 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:40:27 crc kubenswrapper[4930]: E0313 09:40:27.971534 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:40:34 crc kubenswrapper[4930]: I0313 09:40:34.923514 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-575b8b7644-p4hwx" Mar 13 09:40:35 crc kubenswrapper[4930]: I0313 09:40:35.040250 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-77757997f5-wlmc4"] Mar 13 09:40:35 crc kubenswrapper[4930]: I0313 09:40:35.041163 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-77757997f5-wlmc4" podUID="0a8d2ceb-f27f-4e71-9c58-64cb813bd385" containerName="heat-api" containerID="cri-o://1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59" gracePeriod=60 Mar 13 09:40:35 crc kubenswrapper[4930]: I0313 09:40:35.601277 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" Mar 13 09:40:35 crc kubenswrapper[4930]: I0313 09:40:35.709665 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-86b9fc57f7-vjhx9"] Mar 13 09:40:35 crc kubenswrapper[4930]: I0313 09:40:35.710068 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" podUID="e1a0aa4e-26e8-4ff5-9850-df0535655e9e" containerName="heat-cfnapi" containerID="cri-o://3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce" gracePeriod=60 Mar 13 09:40:36 crc kubenswrapper[4930]: I0313 09:40:36.043103 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-2" podUID="bc0f868a-2947-4e5a-a502-7ece4237551c" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.28:5671: connect: connection refused" Mar 13 09:40:36 crc kubenswrapper[4930]: I0313 09:40:36.606636 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" event={"ID":"e68e5869-c983-4ae1-a8c7-d6c4fba11d30","Type":"ContainerStarted","Data":"34918e5889a4c06dab65c18412c4e7c2020827261ef5d63cb3fee7fbc4f4969a"} Mar 13 09:40:36 crc kubenswrapper[4930]: I0313 09:40:36.625852 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" podStartSLOduration=2.984821543 podStartE2EDuration="16.625836658s" podCreationTimestamp="2026-03-13 09:40:20 +0000 UTC" firstStartedPulling="2026-03-13 09:40:21.977910299 +0000 UTC m=+1662.727824976" lastFinishedPulling="2026-03-13 09:40:35.618925414 +0000 UTC m=+1676.368840091" observedRunningTime="2026-03-13 09:40:36.623261413 +0000 UTC m=+1677.373176090" watchObservedRunningTime="2026-03-13 09:40:36.625836658 +0000 UTC m=+1677.375751335" Mar 13 09:40:38 crc kubenswrapper[4930]: I0313 09:40:38.164909 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="2392424a-e626-4305-865e-4817f46a9b65" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.1.29:5671: connect: connection refused" Mar 13 09:40:38 crc kubenswrapper[4930]: I0313 09:40:38.553383 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-77757997f5-wlmc4" podUID="0a8d2ceb-f27f-4e71-9c58-64cb813bd385" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.234:8004/healthcheck\": read tcp 10.217.0.2:54412->10.217.0.234:8004: read: connection reset by peer" Mar 13 09:40:38 crc kubenswrapper[4930]: I0313 09:40:38.970927 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:40:38 crc kubenswrapper[4930]: E0313 09:40:38.971530 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.180228 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.322269 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-config-data-custom\") pod \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.322331 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-public-tls-certs\") pod \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.322421 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-combined-ca-bundle\") pod \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.322537 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-internal-tls-certs\") pod \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.322672 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkbp2\" (UniqueName: \"kubernetes.io/projected/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-kube-api-access-mkbp2\") pod \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.322698 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-config-data\") pod \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\" (UID: \"0a8d2ceb-f27f-4e71-9c58-64cb813bd385\") " Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.329252 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "0a8d2ceb-f27f-4e71-9c58-64cb813bd385" (UID: "0a8d2ceb-f27f-4e71-9c58-64cb813bd385"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.331886 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-kube-api-access-mkbp2" (OuterVolumeSpecName: "kube-api-access-mkbp2") pod "0a8d2ceb-f27f-4e71-9c58-64cb813bd385" (UID: "0a8d2ceb-f27f-4e71-9c58-64cb813bd385"). InnerVolumeSpecName "kube-api-access-mkbp2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.377665 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a8d2ceb-f27f-4e71-9c58-64cb813bd385" (UID: "0a8d2ceb-f27f-4e71-9c58-64cb813bd385"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.391693 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "0a8d2ceb-f27f-4e71-9c58-64cb813bd385" (UID: "0a8d2ceb-f27f-4e71-9c58-64cb813bd385"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.400336 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-config-data" (OuterVolumeSpecName: "config-data") pod "0a8d2ceb-f27f-4e71-9c58-64cb813bd385" (UID: "0a8d2ceb-f27f-4e71-9c58-64cb813bd385"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.411311 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "0a8d2ceb-f27f-4e71-9c58-64cb813bd385" (UID: "0a8d2ceb-f27f-4e71-9c58-64cb813bd385"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.424980 4930 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.425007 4930 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.425018 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.425026 4930 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.425037 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkbp2\" (UniqueName: \"kubernetes.io/projected/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-kube-api-access-mkbp2\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.425047 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0a8d2ceb-f27f-4e71-9c58-64cb813bd385-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.435163 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.628557 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hz4tm\" (UniqueName: \"kubernetes.io/projected/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-kube-api-access-hz4tm\") pod \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.628658 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-combined-ca-bundle\") pod \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.628703 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-public-tls-certs\") pod \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.628792 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-config-data\") pod \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.629032 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-internal-tls-certs\") pod \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.629073 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-config-data-custom\") pod \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\" (UID: \"e1a0aa4e-26e8-4ff5-9850-df0535655e9e\") " Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.631950 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-kube-api-access-hz4tm" (OuterVolumeSpecName: "kube-api-access-hz4tm") pod "e1a0aa4e-26e8-4ff5-9850-df0535655e9e" (UID: "e1a0aa4e-26e8-4ff5-9850-df0535655e9e"). InnerVolumeSpecName "kube-api-access-hz4tm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.632457 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e1a0aa4e-26e8-4ff5-9850-df0535655e9e" (UID: "e1a0aa4e-26e8-4ff5-9850-df0535655e9e"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.637397 4930 generic.go:334] "Generic (PLEG): container finished" podID="0a8d2ceb-f27f-4e71-9c58-64cb813bd385" containerID="1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59" exitCode=0 Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.637489 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-77757997f5-wlmc4" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.637529 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77757997f5-wlmc4" event={"ID":"0a8d2ceb-f27f-4e71-9c58-64cb813bd385","Type":"ContainerDied","Data":"1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59"} Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.637561 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-77757997f5-wlmc4" event={"ID":"0a8d2ceb-f27f-4e71-9c58-64cb813bd385","Type":"ContainerDied","Data":"a20c6ae78f99aae409e30b2ea2567054f5a9075cf0a6f7da5f6ac13dcc528aac"} Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.637578 4930 scope.go:117] "RemoveContainer" containerID="1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.642756 4930 generic.go:334] "Generic (PLEG): container finished" podID="e1a0aa4e-26e8-4ff5-9850-df0535655e9e" containerID="3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce" exitCode=0 Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.642784 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" event={"ID":"e1a0aa4e-26e8-4ff5-9850-df0535655e9e","Type":"ContainerDied","Data":"3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce"} Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.642803 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" event={"ID":"e1a0aa4e-26e8-4ff5-9850-df0535655e9e","Type":"ContainerDied","Data":"01a84cac61dddc07677cb61c72f2195a66659952ef9c47f94e38afb874c51f8e"} Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.642848 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-86b9fc57f7-vjhx9" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.673367 4930 scope.go:117] "RemoveContainer" containerID="1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59" Mar 13 09:40:39 crc kubenswrapper[4930]: E0313 09:40:39.674294 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59\": container with ID starting with 1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59 not found: ID does not exist" containerID="1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.674330 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59"} err="failed to get container status \"1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59\": rpc error: code = NotFound desc = could not find container \"1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59\": container with ID starting with 1c24fdac28b9d2113f3cb4d2f6846da8850ca68b7475266d5967c2cba475aa59 not found: ID does not exist" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.674349 4930 scope.go:117] "RemoveContainer" containerID="3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.686179 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1a0aa4e-26e8-4ff5-9850-df0535655e9e" (UID: "e1a0aa4e-26e8-4ff5-9850-df0535655e9e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.712005 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e1a0aa4e-26e8-4ff5-9850-df0535655e9e" (UID: "e1a0aa4e-26e8-4ff5-9850-df0535655e9e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.716847 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e1a0aa4e-26e8-4ff5-9850-df0535655e9e" (UID: "e1a0aa4e-26e8-4ff5-9850-df0535655e9e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.718991 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-config-data" (OuterVolumeSpecName: "config-data") pod "e1a0aa4e-26e8-4ff5-9850-df0535655e9e" (UID: "e1a0aa4e-26e8-4ff5-9850-df0535655e9e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.732912 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.733174 4930 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.733254 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.733334 4930 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.733425 4930 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.733529 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hz4tm\" (UniqueName: \"kubernetes.io/projected/e1a0aa4e-26e8-4ff5-9850-df0535655e9e-kube-api-access-hz4tm\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.886051 4930 scope.go:117] "RemoveContainer" containerID="3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce" Mar 13 09:40:39 crc kubenswrapper[4930]: E0313 09:40:39.887333 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce\": container with ID starting with 3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce not found: ID does not exist" containerID="3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.887462 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce"} err="failed to get container status \"3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce\": rpc error: code = NotFound desc = could not find container \"3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce\": container with ID starting with 3906d4a0e11645a3c2e3a615d3956b11073c5e71d21ccdda3eb9606b5cadb0ce not found: ID does not exist" Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.891187 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-77757997f5-wlmc4"] Mar 13 09:40:39 crc kubenswrapper[4930]: I0313 09:40:39.903574 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-77757997f5-wlmc4"] Mar 13 09:40:40 crc kubenswrapper[4930]: I0313 09:40:40.011874 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a8d2ceb-f27f-4e71-9c58-64cb813bd385" path="/var/lib/kubelet/pods/0a8d2ceb-f27f-4e71-9c58-64cb813bd385/volumes" Mar 13 09:40:40 crc kubenswrapper[4930]: I0313 09:40:40.012660 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-86b9fc57f7-vjhx9"] Mar 13 09:40:40 crc kubenswrapper[4930]: I0313 09:40:40.037900 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-86b9fc57f7-vjhx9"] Mar 13 09:40:41 crc kubenswrapper[4930]: I0313 09:40:41.119910 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-f9949f6d4-jldnd" Mar 13 09:40:41 crc kubenswrapper[4930]: I0313 09:40:41.172724 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-c6b696d47-c76xv"] Mar 13 09:40:41 crc kubenswrapper[4930]: I0313 09:40:41.172980 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-c6b696d47-c76xv" podUID="f0868712-845b-4803-b7a7-44aaa23ad284" containerName="heat-engine" containerID="cri-o://d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f" gracePeriod=60 Mar 13 09:40:41 crc kubenswrapper[4930]: I0313 09:40:41.984371 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1a0aa4e-26e8-4ff5-9850-df0535655e9e" path="/var/lib/kubelet/pods/e1a0aa4e-26e8-4ff5-9850-df0535655e9e/volumes" Mar 13 09:40:43 crc kubenswrapper[4930]: E0313 09:40:43.490684 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 09:40:43 crc kubenswrapper[4930]: E0313 09:40:43.495841 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 09:40:43 crc kubenswrapper[4930]: E0313 09:40:43.500826 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 09:40:43 crc kubenswrapper[4930]: E0313 09:40:43.500904 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-c6b696d47-c76xv" podUID="f0868712-845b-4803-b7a7-44aaa23ad284" containerName="heat-engine" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.085183 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-dh2pc"] Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.098836 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-dh2pc"] Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.220223 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-4rcdw"] Mar 13 09:40:44 crc kubenswrapper[4930]: E0313 09:40:44.221085 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1a0aa4e-26e8-4ff5-9850-df0535655e9e" containerName="heat-cfnapi" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.221189 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1a0aa4e-26e8-4ff5-9850-df0535655e9e" containerName="heat-cfnapi" Mar 13 09:40:44 crc kubenswrapper[4930]: E0313 09:40:44.221315 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a8d2ceb-f27f-4e71-9c58-64cb813bd385" containerName="heat-api" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.221393 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a8d2ceb-f27f-4e71-9c58-64cb813bd385" containerName="heat-api" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.221801 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a8d2ceb-f27f-4e71-9c58-64cb813bd385" containerName="heat-api" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.221903 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1a0aa4e-26e8-4ff5-9850-df0535655e9e" containerName="heat-cfnapi" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.222955 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.226492 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.232985 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-4rcdw"] Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.244721 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-combined-ca-bundle\") pod \"aodh-db-sync-4rcdw\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.244804 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq265\" (UniqueName: \"kubernetes.io/projected/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-kube-api-access-nq265\") pod \"aodh-db-sync-4rcdw\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.244913 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-config-data\") pod \"aodh-db-sync-4rcdw\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.245064 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-scripts\") pod \"aodh-db-sync-4rcdw\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.346721 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-scripts\") pod \"aodh-db-sync-4rcdw\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.347127 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-combined-ca-bundle\") pod \"aodh-db-sync-4rcdw\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.347180 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq265\" (UniqueName: \"kubernetes.io/projected/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-kube-api-access-nq265\") pod \"aodh-db-sync-4rcdw\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.347284 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-config-data\") pod \"aodh-db-sync-4rcdw\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.366194 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-scripts\") pod \"aodh-db-sync-4rcdw\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.369947 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-config-data\") pod \"aodh-db-sync-4rcdw\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.372106 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq265\" (UniqueName: \"kubernetes.io/projected/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-kube-api-access-nq265\") pod \"aodh-db-sync-4rcdw\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.372314 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-combined-ca-bundle\") pod \"aodh-db-sync-4rcdw\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:44 crc kubenswrapper[4930]: I0313 09:40:44.550444 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:45 crc kubenswrapper[4930]: I0313 09:40:45.077125 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-4rcdw"] Mar 13 09:40:45 crc kubenswrapper[4930]: I0313 09:40:45.730691 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-4rcdw" event={"ID":"18da2b4c-e41d-4e0c-b470-b9651e7f1aae","Type":"ContainerStarted","Data":"1aeba979eb57485fbc63334bc21de3a95d0d339d26f7d4a8fbcec4d1011c238e"} Mar 13 09:40:45 crc kubenswrapper[4930]: I0313 09:40:45.984043 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="17fdbf0c-1335-4369-84b3-7db681759d19" path="/var/lib/kubelet/pods/17fdbf0c-1335-4369-84b3-7db681759d19/volumes" Mar 13 09:40:46 crc kubenswrapper[4930]: I0313 09:40:46.041881 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-2" Mar 13 09:40:46 crc kubenswrapper[4930]: I0313 09:40:46.104888 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 13 09:40:48 crc kubenswrapper[4930]: I0313 09:40:48.163647 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 13 09:40:48 crc kubenswrapper[4930]: I0313 09:40:48.777238 4930 generic.go:334] "Generic (PLEG): container finished" podID="e68e5869-c983-4ae1-a8c7-d6c4fba11d30" containerID="34918e5889a4c06dab65c18412c4e7c2020827261ef5d63cb3fee7fbc4f4969a" exitCode=0 Mar 13 09:40:48 crc kubenswrapper[4930]: I0313 09:40:48.777291 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" event={"ID":"e68e5869-c983-4ae1-a8c7-d6c4fba11d30","Type":"ContainerDied","Data":"34918e5889a4c06dab65c18412c4e7c2020827261ef5d63cb3fee7fbc4f4969a"} Mar 13 09:40:50 crc kubenswrapper[4930]: I0313 09:40:50.972144 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:40:50 crc kubenswrapper[4930]: E0313 09:40:50.972866 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.154794 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-1" podUID="2f69fa32-5b72-4f9b-9176-d43ee011605d" containerName="rabbitmq" containerID="cri-o://8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6" gracePeriod=604795 Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.767706 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.819477 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" event={"ID":"e68e5869-c983-4ae1-a8c7-d6c4fba11d30","Type":"ContainerDied","Data":"d794449ec86f8a8907058a764570206d3641bab10a4709aa32632690394d7c67"} Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.819549 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d794449ec86f8a8907058a764570206d3641bab10a4709aa32632690394d7c67" Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.819615 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-j6tvx" Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.857415 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4lqv\" (UniqueName: \"kubernetes.io/projected/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-kube-api-access-w4lqv\") pod \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.857529 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-ssh-key-openstack-edpm-ipam\") pod \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.857570 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-inventory\") pod \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.857773 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-repo-setup-combined-ca-bundle\") pod \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\" (UID: \"e68e5869-c983-4ae1-a8c7-d6c4fba11d30\") " Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.863399 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "e68e5869-c983-4ae1-a8c7-d6c4fba11d30" (UID: "e68e5869-c983-4ae1-a8c7-d6c4fba11d30"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.864886 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-kube-api-access-w4lqv" (OuterVolumeSpecName: "kube-api-access-w4lqv") pod "e68e5869-c983-4ae1-a8c7-d6c4fba11d30" (UID: "e68e5869-c983-4ae1-a8c7-d6c4fba11d30"). InnerVolumeSpecName "kube-api-access-w4lqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.893425 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-inventory" (OuterVolumeSpecName: "inventory") pod "e68e5869-c983-4ae1-a8c7-d6c4fba11d30" (UID: "e68e5869-c983-4ae1-a8c7-d6c4fba11d30"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.909469 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e68e5869-c983-4ae1-a8c7-d6c4fba11d30" (UID: "e68e5869-c983-4ae1-a8c7-d6c4fba11d30"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.963902 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4lqv\" (UniqueName: \"kubernetes.io/projected/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-kube-api-access-w4lqv\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.963940 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.963955 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:51 crc kubenswrapper[4930]: I0313 09:40:51.963967 4930 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e68e5869-c983-4ae1-a8c7-d6c4fba11d30-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.839546 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-4rcdw" event={"ID":"18da2b4c-e41d-4e0c-b470-b9651e7f1aae","Type":"ContainerStarted","Data":"65b572c3008282bf84784b2e4acabdb2a2c4f34701ed1143700ac0cdaca63253"} Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.912146 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-4rcdw" podStartSLOduration=2.408963605 podStartE2EDuration="8.912120721s" podCreationTimestamp="2026-03-13 09:40:44 +0000 UTC" firstStartedPulling="2026-03-13 09:40:45.072332981 +0000 UTC m=+1685.822247658" lastFinishedPulling="2026-03-13 09:40:51.575490097 +0000 UTC m=+1692.325404774" observedRunningTime="2026-03-13 09:40:52.872763102 +0000 UTC m=+1693.622677769" watchObservedRunningTime="2026-03-13 09:40:52.912120721 +0000 UTC m=+1693.662035398" Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.913898 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn"] Mar 13 09:40:52 crc kubenswrapper[4930]: E0313 09:40:52.914689 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e68e5869-c983-4ae1-a8c7-d6c4fba11d30" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.914718 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="e68e5869-c983-4ae1-a8c7-d6c4fba11d30" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.915047 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="e68e5869-c983-4ae1-a8c7-d6c4fba11d30" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.916123 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.918162 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.918693 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.918868 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.919178 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.930304 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn"] Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.987461 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h56sj\" (UniqueName: \"kubernetes.io/projected/504b0add-0f50-4c3b-9ecc-4a5a297cc400-kube-api-access-h56sj\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-grcrn\" (UID: \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.988846 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/504b0add-0f50-4c3b-9ecc-4a5a297cc400-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-grcrn\" (UID: \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:40:52 crc kubenswrapper[4930]: I0313 09:40:52.988994 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/504b0add-0f50-4c3b-9ecc-4a5a297cc400-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-grcrn\" (UID: \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.094014 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h56sj\" (UniqueName: \"kubernetes.io/projected/504b0add-0f50-4c3b-9ecc-4a5a297cc400-kube-api-access-h56sj\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-grcrn\" (UID: \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.094207 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/504b0add-0f50-4c3b-9ecc-4a5a297cc400-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-grcrn\" (UID: \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.094270 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/504b0add-0f50-4c3b-9ecc-4a5a297cc400-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-grcrn\" (UID: \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.104949 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/504b0add-0f50-4c3b-9ecc-4a5a297cc400-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-grcrn\" (UID: \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.106477 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/504b0add-0f50-4c3b-9ecc-4a5a297cc400-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-grcrn\" (UID: \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.148664 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h56sj\" (UniqueName: \"kubernetes.io/projected/504b0add-0f50-4c3b-9ecc-4a5a297cc400-kube-api-access-h56sj\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-grcrn\" (UID: \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.160039 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-1" podUID="2f69fa32-5b72-4f9b-9176-d43ee011605d" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.137:5671: connect: connection refused" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.260596 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:40:53 crc kubenswrapper[4930]: E0313 09:40:53.489445 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f is running failed: container process not found" containerID="d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 09:40:53 crc kubenswrapper[4930]: E0313 09:40:53.490218 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f is running failed: container process not found" containerID="d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 09:40:53 crc kubenswrapper[4930]: E0313 09:40:53.490488 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f is running failed: container process not found" containerID="d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Mar 13 09:40:53 crc kubenswrapper[4930]: E0313 09:40:53.490535 4930 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f is running failed: container process not found" probeType="Readiness" pod="openstack/heat-engine-c6b696d47-c76xv" podUID="f0868712-845b-4803-b7a7-44aaa23ad284" containerName="heat-engine" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.563598 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.712968 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-config-data-custom\") pod \"f0868712-845b-4803-b7a7-44aaa23ad284\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.713247 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-combined-ca-bundle\") pod \"f0868712-845b-4803-b7a7-44aaa23ad284\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.713602 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-config-data\") pod \"f0868712-845b-4803-b7a7-44aaa23ad284\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.714043 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6p8g8\" (UniqueName: \"kubernetes.io/projected/f0868712-845b-4803-b7a7-44aaa23ad284-kube-api-access-6p8g8\") pod \"f0868712-845b-4803-b7a7-44aaa23ad284\" (UID: \"f0868712-845b-4803-b7a7-44aaa23ad284\") " Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.721698 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0868712-845b-4803-b7a7-44aaa23ad284-kube-api-access-6p8g8" (OuterVolumeSpecName: "kube-api-access-6p8g8") pod "f0868712-845b-4803-b7a7-44aaa23ad284" (UID: "f0868712-845b-4803-b7a7-44aaa23ad284"). InnerVolumeSpecName "kube-api-access-6p8g8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.722882 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f0868712-845b-4803-b7a7-44aaa23ad284" (UID: "f0868712-845b-4803-b7a7-44aaa23ad284"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.761906 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0868712-845b-4803-b7a7-44aaa23ad284" (UID: "f0868712-845b-4803-b7a7-44aaa23ad284"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.791685 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-config-data" (OuterVolumeSpecName: "config-data") pod "f0868712-845b-4803-b7a7-44aaa23ad284" (UID: "f0868712-845b-4803-b7a7-44aaa23ad284"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.816893 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.816922 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.816931 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6p8g8\" (UniqueName: \"kubernetes.io/projected/f0868712-845b-4803-b7a7-44aaa23ad284-kube-api-access-6p8g8\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.816942 4930 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0868712-845b-4803-b7a7-44aaa23ad284-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.854306 4930 generic.go:334] "Generic (PLEG): container finished" podID="f0868712-845b-4803-b7a7-44aaa23ad284" containerID="d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f" exitCode=0 Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.854381 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-c6b696d47-c76xv" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.854466 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-c6b696d47-c76xv" event={"ID":"f0868712-845b-4803-b7a7-44aaa23ad284","Type":"ContainerDied","Data":"d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f"} Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.854506 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-c6b696d47-c76xv" event={"ID":"f0868712-845b-4803-b7a7-44aaa23ad284","Type":"ContainerDied","Data":"47b3c8d3318c0e56add86aca51d17f8bc3f46e4536e054a342375d27f53241f4"} Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.854537 4930 scope.go:117] "RemoveContainer" containerID="d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.917604 4930 scope.go:117] "RemoveContainer" containerID="d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f" Mar 13 09:40:53 crc kubenswrapper[4930]: E0313 09:40:53.919274 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f\": container with ID starting with d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f not found: ID does not exist" containerID="d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.919305 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f"} err="failed to get container status \"d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f\": rpc error: code = NotFound desc = could not find container \"d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f\": container with ID starting with d8b842e2545c566d411c3e6c3ea28bccc0b74f4efded0770142a1f064499736f not found: ID does not exist" Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.920490 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-c6b696d47-c76xv"] Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.931212 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-c6b696d47-c76xv"] Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.951157 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn"] Mar 13 09:40:53 crc kubenswrapper[4930]: I0313 09:40:53.986171 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0868712-845b-4803-b7a7-44aaa23ad284" path="/var/lib/kubelet/pods/f0868712-845b-4803-b7a7-44aaa23ad284/volumes" Mar 13 09:40:54 crc kubenswrapper[4930]: I0313 09:40:54.869393 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" event={"ID":"504b0add-0f50-4c3b-9ecc-4a5a297cc400","Type":"ContainerStarted","Data":"d4943cbaaf5afce0d2a4f7f6fac3166b244fa7b6c1c32b5c6dabcf1cbf26b636"} Mar 13 09:40:55 crc kubenswrapper[4930]: I0313 09:40:55.894672 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" event={"ID":"504b0add-0f50-4c3b-9ecc-4a5a297cc400","Type":"ContainerStarted","Data":"5bd48a5515738bba027d810ebe83e0d53a0d8974d3e59a0d30350f77a658a1fe"} Mar 13 09:40:55 crc kubenswrapper[4930]: I0313 09:40:55.897136 4930 generic.go:334] "Generic (PLEG): container finished" podID="18da2b4c-e41d-4e0c-b470-b9651e7f1aae" containerID="65b572c3008282bf84784b2e4acabdb2a2c4f34701ed1143700ac0cdaca63253" exitCode=0 Mar 13 09:40:55 crc kubenswrapper[4930]: I0313 09:40:55.897181 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-4rcdw" event={"ID":"18da2b4c-e41d-4e0c-b470-b9651e7f1aae","Type":"ContainerDied","Data":"65b572c3008282bf84784b2e4acabdb2a2c4f34701ed1143700ac0cdaca63253"} Mar 13 09:40:55 crc kubenswrapper[4930]: I0313 09:40:55.911918 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" podStartSLOduration=3.397737117 podStartE2EDuration="3.911901721s" podCreationTimestamp="2026-03-13 09:40:52 +0000 UTC" firstStartedPulling="2026-03-13 09:40:53.957642776 +0000 UTC m=+1694.707557453" lastFinishedPulling="2026-03-13 09:40:54.47180737 +0000 UTC m=+1695.221722057" observedRunningTime="2026-03-13 09:40:55.908611817 +0000 UTC m=+1696.658526494" watchObservedRunningTime="2026-03-13 09:40:55.911901721 +0000 UTC m=+1696.661816398" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.355662 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.404887 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-scripts\") pod \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.404937 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-combined-ca-bundle\") pod \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.405231 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-config-data\") pod \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.405323 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nq265\" (UniqueName: \"kubernetes.io/projected/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-kube-api-access-nq265\") pod \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\" (UID: \"18da2b4c-e41d-4e0c-b470-b9651e7f1aae\") " Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.415703 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-scripts" (OuterVolumeSpecName: "scripts") pod "18da2b4c-e41d-4e0c-b470-b9651e7f1aae" (UID: "18da2b4c-e41d-4e0c-b470-b9651e7f1aae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.430677 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-kube-api-access-nq265" (OuterVolumeSpecName: "kube-api-access-nq265") pod "18da2b4c-e41d-4e0c-b470-b9651e7f1aae" (UID: "18da2b4c-e41d-4e0c-b470-b9651e7f1aae"). InnerVolumeSpecName "kube-api-access-nq265". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.442100 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-config-data" (OuterVolumeSpecName: "config-data") pod "18da2b4c-e41d-4e0c-b470-b9651e7f1aae" (UID: "18da2b4c-e41d-4e0c-b470-b9651e7f1aae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.463412 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "18da2b4c-e41d-4e0c-b470-b9651e7f1aae" (UID: "18da2b4c-e41d-4e0c-b470-b9651e7f1aae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.508160 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.508185 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nq265\" (UniqueName: \"kubernetes.io/projected/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-kube-api-access-nq265\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.508197 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.508206 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/18da2b4c-e41d-4e0c-b470-b9651e7f1aae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.896732 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.956307 4930 generic.go:334] "Generic (PLEG): container finished" podID="2f69fa32-5b72-4f9b-9176-d43ee011605d" containerID="8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6" exitCode=0 Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.956364 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"2f69fa32-5b72-4f9b-9176-d43ee011605d","Type":"ContainerDied","Data":"8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6"} Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.956378 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.956426 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"2f69fa32-5b72-4f9b-9176-d43ee011605d","Type":"ContainerDied","Data":"63f2bb7080932271c381d7af1c9b19e5a24762dd7d2af49248b5c79d9709c96b"} Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.956519 4930 scope.go:117] "RemoveContainer" containerID="8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6" Mar 13 09:40:57 crc kubenswrapper[4930]: I0313 09:40:57.974186 4930 generic.go:334] "Generic (PLEG): container finished" podID="504b0add-0f50-4c3b-9ecc-4a5a297cc400" containerID="5bd48a5515738bba027d810ebe83e0d53a0d8974d3e59a0d30350f77a658a1fe" exitCode=0 Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.004758 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-4rcdw" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.008742 4930 scope.go:117] "RemoveContainer" containerID="f5bb5cadafeb7d370421ec827fb7acdb9c3d0c47530810cf1b6e9e27c93891de" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.021456 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" event={"ID":"504b0add-0f50-4c3b-9ecc-4a5a297cc400","Type":"ContainerDied","Data":"5bd48a5515738bba027d810ebe83e0d53a0d8974d3e59a0d30350f77a658a1fe"} Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.021507 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-4rcdw" event={"ID":"18da2b4c-e41d-4e0c-b470-b9651e7f1aae","Type":"ContainerDied","Data":"1aeba979eb57485fbc63334bc21de3a95d0d339d26f7d4a8fbcec4d1011c238e"} Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.021526 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1aeba979eb57485fbc63334bc21de3a95d0d339d26f7d4a8fbcec4d1011c238e" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.034380 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f69fa32-5b72-4f9b-9176-d43ee011605d-erlang-cookie-secret\") pod \"2f69fa32-5b72-4f9b-9176-d43ee011605d\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.034454 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-config-data\") pod \"2f69fa32-5b72-4f9b-9176-d43ee011605d\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.034476 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-erlang-cookie\") pod \"2f69fa32-5b72-4f9b-9176-d43ee011605d\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.034613 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-tls\") pod \"2f69fa32-5b72-4f9b-9176-d43ee011605d\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.034688 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f69fa32-5b72-4f9b-9176-d43ee011605d-pod-info\") pod \"2f69fa32-5b72-4f9b-9176-d43ee011605d\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.034769 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-plugins\") pod \"2f69fa32-5b72-4f9b-9176-d43ee011605d\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.034820 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-plugins-conf\") pod \"2f69fa32-5b72-4f9b-9176-d43ee011605d\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.035027 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2f69fa32-5b72-4f9b-9176-d43ee011605d" (UID: "2f69fa32-5b72-4f9b-9176-d43ee011605d"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.042010 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2f69fa32-5b72-4f9b-9176-d43ee011605d" (UID: "2f69fa32-5b72-4f9b-9176-d43ee011605d"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.044006 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\") pod \"2f69fa32-5b72-4f9b-9176-d43ee011605d\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.044072 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dchqq\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-kube-api-access-dchqq\") pod \"2f69fa32-5b72-4f9b-9176-d43ee011605d\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.044163 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-server-conf\") pod \"2f69fa32-5b72-4f9b-9176-d43ee011605d\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.044214 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-confd\") pod \"2f69fa32-5b72-4f9b-9176-d43ee011605d\" (UID: \"2f69fa32-5b72-4f9b-9176-d43ee011605d\") " Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.045297 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2f69fa32-5b72-4f9b-9176-d43ee011605d" (UID: "2f69fa32-5b72-4f9b-9176-d43ee011605d"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.046155 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.046606 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.046955 4930 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.044421 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2f69fa32-5b72-4f9b-9176-d43ee011605d-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2f69fa32-5b72-4f9b-9176-d43ee011605d" (UID: "2f69fa32-5b72-4f9b-9176-d43ee011605d"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.060958 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2f69fa32-5b72-4f9b-9176-d43ee011605d" (UID: "2f69fa32-5b72-4f9b-9176-d43ee011605d"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.064253 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2f69fa32-5b72-4f9b-9176-d43ee011605d-pod-info" (OuterVolumeSpecName: "pod-info") pod "2f69fa32-5b72-4f9b-9176-d43ee011605d" (UID: "2f69fa32-5b72-4f9b-9176-d43ee011605d"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.068634 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-kube-api-access-dchqq" (OuterVolumeSpecName: "kube-api-access-dchqq") pod "2f69fa32-5b72-4f9b-9176-d43ee011605d" (UID: "2f69fa32-5b72-4f9b-9176-d43ee011605d"). InnerVolumeSpecName "kube-api-access-dchqq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.114461 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-config-data" (OuterVolumeSpecName: "config-data") pod "2f69fa32-5b72-4f9b-9176-d43ee011605d" (UID: "2f69fa32-5b72-4f9b-9176-d43ee011605d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.149116 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.149146 4930 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2f69fa32-5b72-4f9b-9176-d43ee011605d-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.149157 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dchqq\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-kube-api-access-dchqq\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.149168 4930 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2f69fa32-5b72-4f9b-9176-d43ee011605d-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.149176 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.156252 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8" (OuterVolumeSpecName: "persistence") pod "2f69fa32-5b72-4f9b-9176-d43ee011605d" (UID: "2f69fa32-5b72-4f9b-9176-d43ee011605d"). InnerVolumeSpecName "pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.168476 4930 scope.go:117] "RemoveContainer" containerID="8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6" Mar 13 09:40:58 crc kubenswrapper[4930]: E0313 09:40:58.169800 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6\": container with ID starting with 8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6 not found: ID does not exist" containerID="8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.169835 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6"} err="failed to get container status \"8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6\": rpc error: code = NotFound desc = could not find container \"8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6\": container with ID starting with 8b374f58f240a758b23aec80e42209050091cb204da5486ed3a3463dd2d555c6 not found: ID does not exist" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.169859 4930 scope.go:117] "RemoveContainer" containerID="f5bb5cadafeb7d370421ec827fb7acdb9c3d0c47530810cf1b6e9e27c93891de" Mar 13 09:40:58 crc kubenswrapper[4930]: E0313 09:40:58.170738 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5bb5cadafeb7d370421ec827fb7acdb9c3d0c47530810cf1b6e9e27c93891de\": container with ID starting with f5bb5cadafeb7d370421ec827fb7acdb9c3d0c47530810cf1b6e9e27c93891de not found: ID does not exist" containerID="f5bb5cadafeb7d370421ec827fb7acdb9c3d0c47530810cf1b6e9e27c93891de" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.170769 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5bb5cadafeb7d370421ec827fb7acdb9c3d0c47530810cf1b6e9e27c93891de"} err="failed to get container status \"f5bb5cadafeb7d370421ec827fb7acdb9c3d0c47530810cf1b6e9e27c93891de\": rpc error: code = NotFound desc = could not find container \"f5bb5cadafeb7d370421ec827fb7acdb9c3d0c47530810cf1b6e9e27c93891de\": container with ID starting with f5bb5cadafeb7d370421ec827fb7acdb9c3d0c47530810cf1b6e9e27c93891de not found: ID does not exist" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.185098 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-server-conf" (OuterVolumeSpecName: "server-conf") pod "2f69fa32-5b72-4f9b-9176-d43ee011605d" (UID: "2f69fa32-5b72-4f9b-9176-d43ee011605d"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.239006 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.239283 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-api" containerID="cri-o://da5c96bf237888298b3a1b2948912a0b21740180ff441606d8c15b2a4abdb6d9" gracePeriod=30 Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.239354 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-listener" containerID="cri-o://ab3db3a80af457bd7b4a8dae8e2548f200e1ae5a50c6bdc6cecd064f95934dee" gracePeriod=30 Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.239405 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-notifier" containerID="cri-o://6750b989b595b6420d94d72e5e2309c5a1199f5e3192788cb02e421aef4949f6" gracePeriod=30 Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.239392 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-evaluator" containerID="cri-o://38c3ee5b46bba573a9741588b13094f9f6b96ca6958a7d7b5bb9367858bdfd17" gracePeriod=30 Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.251024 4930 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\") on node \"crc\" " Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.251056 4930 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2f69fa32-5b72-4f9b-9176-d43ee011605d-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.267258 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2f69fa32-5b72-4f9b-9176-d43ee011605d" (UID: "2f69fa32-5b72-4f9b-9176-d43ee011605d"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.316009 4930 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.316176 4930 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8") on node "crc" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.352889 4930 reconciler_common.go:293] "Volume detached for volume \"pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.352938 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2f69fa32-5b72-4f9b-9176-d43ee011605d-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.595589 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.611251 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.623941 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-1"] Mar 13 09:40:58 crc kubenswrapper[4930]: E0313 09:40:58.624462 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f69fa32-5b72-4f9b-9176-d43ee011605d" containerName="rabbitmq" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.624485 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f69fa32-5b72-4f9b-9176-d43ee011605d" containerName="rabbitmq" Mar 13 09:40:58 crc kubenswrapper[4930]: E0313 09:40:58.624509 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2f69fa32-5b72-4f9b-9176-d43ee011605d" containerName="setup-container" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.624516 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2f69fa32-5b72-4f9b-9176-d43ee011605d" containerName="setup-container" Mar 13 09:40:58 crc kubenswrapper[4930]: E0313 09:40:58.624552 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0868712-845b-4803-b7a7-44aaa23ad284" containerName="heat-engine" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.624562 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0868712-845b-4803-b7a7-44aaa23ad284" containerName="heat-engine" Mar 13 09:40:58 crc kubenswrapper[4930]: E0313 09:40:58.624591 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18da2b4c-e41d-4e0c-b470-b9651e7f1aae" containerName="aodh-db-sync" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.624603 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="18da2b4c-e41d-4e0c-b470-b9651e7f1aae" containerName="aodh-db-sync" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.624832 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="2f69fa32-5b72-4f9b-9176-d43ee011605d" containerName="rabbitmq" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.624866 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0868712-845b-4803-b7a7-44aaa23ad284" containerName="heat-engine" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.624881 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="18da2b4c-e41d-4e0c-b470-b9651e7f1aae" containerName="aodh-db-sync" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.654662 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.654822 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.775700 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ab28cafc-95dc-4ace-a257-a428df08e6ef-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.775764 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ab28cafc-95dc-4ace-a257-a428df08e6ef-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.777234 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ab28cafc-95dc-4ace-a257-a428df08e6ef-pod-info\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.777295 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nx2dc\" (UniqueName: \"kubernetes.io/projected/ab28cafc-95dc-4ace-a257-a428df08e6ef-kube-api-access-nx2dc\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.777362 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ab28cafc-95dc-4ace-a257-a428df08e6ef-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.777611 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ab28cafc-95dc-4ace-a257-a428df08e6ef-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.777686 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ab28cafc-95dc-4ace-a257-a428df08e6ef-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.777733 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ab28cafc-95dc-4ace-a257-a428df08e6ef-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.777777 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.777807 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ab28cafc-95dc-4ace-a257-a428df08e6ef-server-conf\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.777835 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab28cafc-95dc-4ace-a257-a428df08e6ef-config-data\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.879988 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ab28cafc-95dc-4ace-a257-a428df08e6ef-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.880750 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.880795 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ab28cafc-95dc-4ace-a257-a428df08e6ef-server-conf\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.880836 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab28cafc-95dc-4ace-a257-a428df08e6ef-config-data\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.880928 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ab28cafc-95dc-4ace-a257-a428df08e6ef-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.880971 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ab28cafc-95dc-4ace-a257-a428df08e6ef-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.880995 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ab28cafc-95dc-4ace-a257-a428df08e6ef-pod-info\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.881032 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nx2dc\" (UniqueName: \"kubernetes.io/projected/ab28cafc-95dc-4ace-a257-a428df08e6ef-kube-api-access-nx2dc\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.881069 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ab28cafc-95dc-4ace-a257-a428df08e6ef-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.881232 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ab28cafc-95dc-4ace-a257-a428df08e6ef-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.881294 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ab28cafc-95dc-4ace-a257-a428df08e6ef-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.881760 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/ab28cafc-95dc-4ace-a257-a428df08e6ef-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.881931 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/ab28cafc-95dc-4ace-a257-a428df08e6ef-rabbitmq-plugins\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.883826 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/ab28cafc-95dc-4ace-a257-a428df08e6ef-server-conf\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.887055 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ab28cafc-95dc-4ace-a257-a428df08e6ef-config-data\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.890726 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/ab28cafc-95dc-4ace-a257-a428df08e6ef-plugins-conf\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.891465 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/ab28cafc-95dc-4ace-a257-a428df08e6ef-rabbitmq-confd\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.891896 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/ab28cafc-95dc-4ace-a257-a428df08e6ef-pod-info\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.895021 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/ab28cafc-95dc-4ace-a257-a428df08e6ef-erlang-cookie-secret\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.896200 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.896252 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/973977d077043ef8fddec8a941dafd7e9bdafd407131659801d4835fea31a670/globalmount\"" pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.911017 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/ab28cafc-95dc-4ace-a257-a428df08e6ef-rabbitmq-tls\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:58 crc kubenswrapper[4930]: I0313 09:40:58.943696 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nx2dc\" (UniqueName: \"kubernetes.io/projected/ab28cafc-95dc-4ace-a257-a428df08e6ef-kube-api-access-nx2dc\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:59 crc kubenswrapper[4930]: I0313 09:40:59.019532 4930 generic.go:334] "Generic (PLEG): container finished" podID="d5419bb0-4585-43ee-9130-234ce8731cde" containerID="38c3ee5b46bba573a9741588b13094f9f6b96ca6958a7d7b5bb9367858bdfd17" exitCode=0 Mar 13 09:40:59 crc kubenswrapper[4930]: I0313 09:40:59.019571 4930 generic.go:334] "Generic (PLEG): container finished" podID="d5419bb0-4585-43ee-9130-234ce8731cde" containerID="da5c96bf237888298b3a1b2948912a0b21740180ff441606d8c15b2a4abdb6d9" exitCode=0 Mar 13 09:40:59 crc kubenswrapper[4930]: I0313 09:40:59.019610 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d5419bb0-4585-43ee-9130-234ce8731cde","Type":"ContainerDied","Data":"38c3ee5b46bba573a9741588b13094f9f6b96ca6958a7d7b5bb9367858bdfd17"} Mar 13 09:40:59 crc kubenswrapper[4930]: I0313 09:40:59.019645 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d5419bb0-4585-43ee-9130-234ce8731cde","Type":"ContainerDied","Data":"da5c96bf237888298b3a1b2948912a0b21740180ff441606d8c15b2a4abdb6d9"} Mar 13 09:40:59 crc kubenswrapper[4930]: I0313 09:40:59.127752 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-6f9a8f8e-3484-429a-ac81-87539d43b6e8\") pod \"rabbitmq-server-1\" (UID: \"ab28cafc-95dc-4ace-a257-a428df08e6ef\") " pod="openstack/rabbitmq-server-1" Mar 13 09:40:59 crc kubenswrapper[4930]: I0313 09:40:59.284990 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-1" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:40:59.541767 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:40:59.600219 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/504b0add-0f50-4c3b-9ecc-4a5a297cc400-inventory\") pod \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\" (UID: \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\") " Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:40:59.600285 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h56sj\" (UniqueName: \"kubernetes.io/projected/504b0add-0f50-4c3b-9ecc-4a5a297cc400-kube-api-access-h56sj\") pod \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\" (UID: \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\") " Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:40:59.600589 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/504b0add-0f50-4c3b-9ecc-4a5a297cc400-ssh-key-openstack-edpm-ipam\") pod \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\" (UID: \"504b0add-0f50-4c3b-9ecc-4a5a297cc400\") " Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:40:59.622664 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/504b0add-0f50-4c3b-9ecc-4a5a297cc400-kube-api-access-h56sj" (OuterVolumeSpecName: "kube-api-access-h56sj") pod "504b0add-0f50-4c3b-9ecc-4a5a297cc400" (UID: "504b0add-0f50-4c3b-9ecc-4a5a297cc400"). InnerVolumeSpecName "kube-api-access-h56sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:40:59.641949 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/504b0add-0f50-4c3b-9ecc-4a5a297cc400-inventory" (OuterVolumeSpecName: "inventory") pod "504b0add-0f50-4c3b-9ecc-4a5a297cc400" (UID: "504b0add-0f50-4c3b-9ecc-4a5a297cc400"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:40:59.655420 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/504b0add-0f50-4c3b-9ecc-4a5a297cc400-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "504b0add-0f50-4c3b-9ecc-4a5a297cc400" (UID: "504b0add-0f50-4c3b-9ecc-4a5a297cc400"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:40:59.703254 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/504b0add-0f50-4c3b-9ecc-4a5a297cc400-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:40:59.703280 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/504b0add-0f50-4c3b-9ecc-4a5a297cc400-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:40:59.703290 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h56sj\" (UniqueName: \"kubernetes.io/projected/504b0add-0f50-4c3b-9ecc-4a5a297cc400-kube-api-access-h56sj\") on node \"crc\" DevicePath \"\"" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:40:59.989029 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2f69fa32-5b72-4f9b-9176-d43ee011605d" path="/var/lib/kubelet/pods/2f69fa32-5b72-4f9b-9176-d43ee011605d/volumes" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.035884 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" event={"ID":"504b0add-0f50-4c3b-9ecc-4a5a297cc400","Type":"ContainerDied","Data":"d4943cbaaf5afce0d2a4f7f6fac3166b244fa7b6c1c32b5c6dabcf1cbf26b636"} Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.035919 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-grcrn" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.035922 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4943cbaaf5afce0d2a4f7f6fac3166b244fa7b6c1c32b5c6dabcf1cbf26b636" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.197542 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s"] Mar 13 09:41:04 crc kubenswrapper[4930]: E0313 09:41:00.198227 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="504b0add-0f50-4c3b-9ecc-4a5a297cc400" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.198247 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="504b0add-0f50-4c3b-9ecc-4a5a297cc400" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.198648 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="504b0add-0f50-4c3b-9ecc-4a5a297cc400" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.199725 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.201578 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.201680 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.202401 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.209495 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s"] Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.215468 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.316982 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.317073 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.317107 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8rkt\" (UniqueName: \"kubernetes.io/projected/1c483489-be42-4344-8485-7959520070cb-kube-api-access-j8rkt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.317153 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.418959 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.419026 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8rkt\" (UniqueName: \"kubernetes.io/projected/1c483489-be42-4344-8485-7959520070cb-kube-api-access-j8rkt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.419085 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.419307 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.423357 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.423708 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.423822 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.443065 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8rkt\" (UniqueName: \"kubernetes.io/projected/1c483489-be42-4344-8485-7959520070cb-kube-api-access-j8rkt\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:00.521530 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:04.087246 4930 generic.go:334] "Generic (PLEG): container finished" podID="d5419bb0-4585-43ee-9130-234ce8731cde" containerID="ab3db3a80af457bd7b4a8dae8e2548f200e1ae5a50c6bdc6cecd064f95934dee" exitCode=0 Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:04.087921 4930 generic.go:334] "Generic (PLEG): container finished" podID="d5419bb0-4585-43ee-9130-234ce8731cde" containerID="6750b989b595b6420d94d72e5e2309c5a1199f5e3192788cb02e421aef4949f6" exitCode=0 Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:04.087310 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d5419bb0-4585-43ee-9130-234ce8731cde","Type":"ContainerDied","Data":"ab3db3a80af457bd7b4a8dae8e2548f200e1ae5a50c6bdc6cecd064f95934dee"} Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:04.087958 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d5419bb0-4585-43ee-9130-234ce8731cde","Type":"ContainerDied","Data":"6750b989b595b6420d94d72e5e2309c5a1199f5e3192788cb02e421aef4949f6"} Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:04.503741 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-1"] Mar 13 09:41:04 crc kubenswrapper[4930]: I0313 09:41:04.643784 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s"] Mar 13 09:41:05 crc kubenswrapper[4930]: I0313 09:41:05.101315 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" event={"ID":"1c483489-be42-4344-8485-7959520070cb","Type":"ContainerStarted","Data":"37d45f544ef16806c6978c45044592f8951c838b0651b98d7e66103fb7a2e920"} Mar 13 09:41:05 crc kubenswrapper[4930]: I0313 09:41:05.104808 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"ab28cafc-95dc-4ace-a257-a428df08e6ef","Type":"ContainerStarted","Data":"c1f2a917eaf15995475cfc45749b55132f13177a7943497cdc900e1d1e9702c3"} Mar 13 09:41:05 crc kubenswrapper[4930]: I0313 09:41:05.912671 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 09:41:05 crc kubenswrapper[4930]: I0313 09:41:05.977466 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:41:05 crc kubenswrapper[4930]: E0313 09:41:05.978168 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:41:05 crc kubenswrapper[4930]: I0313 09:41:05.991762 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-scripts\") pod \"d5419bb0-4585-43ee-9130-234ce8731cde\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " Mar 13 09:41:05 crc kubenswrapper[4930]: I0313 09:41:05.991895 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-combined-ca-bundle\") pod \"d5419bb0-4585-43ee-9130-234ce8731cde\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " Mar 13 09:41:05 crc kubenswrapper[4930]: I0313 09:41:05.992044 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-internal-tls-certs\") pod \"d5419bb0-4585-43ee-9130-234ce8731cde\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " Mar 13 09:41:05 crc kubenswrapper[4930]: I0313 09:41:05.992106 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jjswq\" (UniqueName: \"kubernetes.io/projected/d5419bb0-4585-43ee-9130-234ce8731cde-kube-api-access-jjswq\") pod \"d5419bb0-4585-43ee-9130-234ce8731cde\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " Mar 13 09:41:05 crc kubenswrapper[4930]: I0313 09:41:05.992253 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-public-tls-certs\") pod \"d5419bb0-4585-43ee-9130-234ce8731cde\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " Mar 13 09:41:05 crc kubenswrapper[4930]: I0313 09:41:05.992341 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-config-data\") pod \"d5419bb0-4585-43ee-9130-234ce8731cde\" (UID: \"d5419bb0-4585-43ee-9130-234ce8731cde\") " Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.124263 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d5419bb0-4585-43ee-9130-234ce8731cde","Type":"ContainerDied","Data":"cfbb96be6f9e279947c86fe9f71b554e91492cdeba927b9693f798fdcd663620"} Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.124314 4930 scope.go:117] "RemoveContainer" containerID="ab3db3a80af457bd7b4a8dae8e2548f200e1ae5a50c6bdc6cecd064f95934dee" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.124517 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.166935 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5419bb0-4585-43ee-9130-234ce8731cde-kube-api-access-jjswq" (OuterVolumeSpecName: "kube-api-access-jjswq") pod "d5419bb0-4585-43ee-9130-234ce8731cde" (UID: "d5419bb0-4585-43ee-9130-234ce8731cde"). InnerVolumeSpecName "kube-api-access-jjswq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.167198 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-scripts" (OuterVolumeSpecName: "scripts") pod "d5419bb0-4585-43ee-9130-234ce8731cde" (UID: "d5419bb0-4585-43ee-9130-234ce8731cde"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.207877 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.207918 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jjswq\" (UniqueName: \"kubernetes.io/projected/d5419bb0-4585-43ee-9130-234ce8731cde-kube-api-access-jjswq\") on node \"crc\" DevicePath \"\"" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.219624 4930 scope.go:117] "RemoveContainer" containerID="6750b989b595b6420d94d72e5e2309c5a1199f5e3192788cb02e421aef4949f6" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.228190 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d5419bb0-4585-43ee-9130-234ce8731cde" (UID: "d5419bb0-4585-43ee-9130-234ce8731cde"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.283814 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "d5419bb0-4585-43ee-9130-234ce8731cde" (UID: "d5419bb0-4585-43ee-9130-234ce8731cde"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.310545 4930 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.310586 4930 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.333617 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d5419bb0-4585-43ee-9130-234ce8731cde" (UID: "d5419bb0-4585-43ee-9130-234ce8731cde"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.356793 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-config-data" (OuterVolumeSpecName: "config-data") pod "d5419bb0-4585-43ee-9130-234ce8731cde" (UID: "d5419bb0-4585-43ee-9130-234ce8731cde"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.412574 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.412603 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d5419bb0-4585-43ee-9130-234ce8731cde-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.483663 4930 scope.go:117] "RemoveContainer" containerID="38c3ee5b46bba573a9741588b13094f9f6b96ca6958a7d7b5bb9367858bdfd17" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.496025 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.525224 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.537276 4930 scope.go:117] "RemoveContainer" containerID="da5c96bf237888298b3a1b2948912a0b21740180ff441606d8c15b2a4abdb6d9" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.559877 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Mar 13 09:41:06 crc kubenswrapper[4930]: E0313 09:41:06.560425 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-notifier" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.560458 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-notifier" Mar 13 09:41:06 crc kubenswrapper[4930]: E0313 09:41:06.560483 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-listener" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.560489 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-listener" Mar 13 09:41:06 crc kubenswrapper[4930]: E0313 09:41:06.560496 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-evaluator" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.560502 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-evaluator" Mar 13 09:41:06 crc kubenswrapper[4930]: E0313 09:41:06.560508 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-api" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.560514 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-api" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.560722 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-notifier" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.560739 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-listener" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.560751 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-api" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.560770 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" containerName="aodh-evaluator" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.563027 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.565283 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-dc4s9" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.565555 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.565880 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.566038 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.567145 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.583846 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.719768 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-internal-tls-certs\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.720019 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-public-tls-certs\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.720161 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-scripts\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.720269 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.720349 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx9t2\" (UniqueName: \"kubernetes.io/projected/d473ca87-fdff-43bd-bc47-1c06e452b085-kube-api-access-gx9t2\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.720387 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-config-data\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.823056 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-internal-tls-certs\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.823198 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-public-tls-certs\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.823270 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-scripts\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.823353 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.823951 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx9t2\" (UniqueName: \"kubernetes.io/projected/d473ca87-fdff-43bd-bc47-1c06e452b085-kube-api-access-gx9t2\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.823989 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-config-data\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.828988 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-internal-tls-certs\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.831978 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-scripts\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.844067 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-combined-ca-bundle\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.844584 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-public-tls-certs\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.845385 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d473ca87-fdff-43bd-bc47-1c06e452b085-config-data\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.848992 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx9t2\" (UniqueName: \"kubernetes.io/projected/d473ca87-fdff-43bd-bc47-1c06e452b085-kube-api-access-gx9t2\") pod \"aodh-0\" (UID: \"d473ca87-fdff-43bd-bc47-1c06e452b085\") " pod="openstack/aodh-0" Mar 13 09:41:06 crc kubenswrapper[4930]: I0313 09:41:06.897002 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Mar 13 09:41:07 crc kubenswrapper[4930]: I0313 09:41:07.149822 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" event={"ID":"1c483489-be42-4344-8485-7959520070cb","Type":"ContainerStarted","Data":"fa64ce7da113407390ca4c267e066263118e71079a856424835f34cb916df480"} Mar 13 09:41:07 crc kubenswrapper[4930]: I0313 09:41:07.152167 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"ab28cafc-95dc-4ace-a257-a428df08e6ef","Type":"ContainerStarted","Data":"0201f8ad55880f3efec8395f9d81bb069ea054bd8cb791509d996e8a00904610"} Mar 13 09:41:07 crc kubenswrapper[4930]: I0313 09:41:07.194664 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" podStartSLOduration=5.620403194 podStartE2EDuration="7.194643782s" podCreationTimestamp="2026-03-13 09:41:00 +0000 UTC" firstStartedPulling="2026-03-13 09:41:04.64533924 +0000 UTC m=+1705.395253917" lastFinishedPulling="2026-03-13 09:41:06.219579828 +0000 UTC m=+1706.969494505" observedRunningTime="2026-03-13 09:41:07.165194395 +0000 UTC m=+1707.915109072" watchObservedRunningTime="2026-03-13 09:41:07.194643782 +0000 UTC m=+1707.944558459" Mar 13 09:41:07 crc kubenswrapper[4930]: I0313 09:41:07.457548 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Mar 13 09:41:07 crc kubenswrapper[4930]: I0313 09:41:07.986182 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5419bb0-4585-43ee-9130-234ce8731cde" path="/var/lib/kubelet/pods/d5419bb0-4585-43ee-9130-234ce8731cde/volumes" Mar 13 09:41:08 crc kubenswrapper[4930]: I0313 09:41:08.169732 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d473ca87-fdff-43bd-bc47-1c06e452b085","Type":"ContainerStarted","Data":"1bf88df232c44a3cb31c29989ffd5761f707c3c3ad807a7c55515cd9f4a8177b"} Mar 13 09:41:08 crc kubenswrapper[4930]: I0313 09:41:08.169962 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d473ca87-fdff-43bd-bc47-1c06e452b085","Type":"ContainerStarted","Data":"201efe0384e1c52121a5a16f31b155c736a18e5baffe2492bba6a9e479147704"} Mar 13 09:41:10 crc kubenswrapper[4930]: I0313 09:41:10.196583 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d473ca87-fdff-43bd-bc47-1c06e452b085","Type":"ContainerStarted","Data":"824c8f9c5e3851c12fdc7deb38b9a32eaa1a1a51876e563302c102e25cd8aad2"} Mar 13 09:41:11 crc kubenswrapper[4930]: I0313 09:41:11.211124 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d473ca87-fdff-43bd-bc47-1c06e452b085","Type":"ContainerStarted","Data":"b79e070f7c35a2970bab0850c699cec9e358077be7b47e5fb030baa2f76cc5fd"} Mar 13 09:41:13 crc kubenswrapper[4930]: I0313 09:41:13.244182 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"d473ca87-fdff-43bd-bc47-1c06e452b085","Type":"ContainerStarted","Data":"d56bcc3951c5baf7802f99aa2cdebee51c7db4eceae91c9f0c54d2e3fa164559"} Mar 13 09:41:13 crc kubenswrapper[4930]: I0313 09:41:13.270344 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.699412197 podStartE2EDuration="7.270323305s" podCreationTimestamp="2026-03-13 09:41:06 +0000 UTC" firstStartedPulling="2026-03-13 09:41:07.475491133 +0000 UTC m=+1708.225405810" lastFinishedPulling="2026-03-13 09:41:12.046402241 +0000 UTC m=+1712.796316918" observedRunningTime="2026-03-13 09:41:13.267566765 +0000 UTC m=+1714.017481442" watchObservedRunningTime="2026-03-13 09:41:13.270323305 +0000 UTC m=+1714.020237982" Mar 13 09:41:19 crc kubenswrapper[4930]: I0313 09:41:19.979895 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:41:19 crc kubenswrapper[4930]: E0313 09:41:19.980788 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:41:27 crc kubenswrapper[4930]: I0313 09:41:27.059407 4930 scope.go:117] "RemoveContainer" containerID="6e35cb5389d90f31f41ef51b7c6a85b9edc8e6f4e042fbf370a1bba9fd1433f9" Mar 13 09:41:27 crc kubenswrapper[4930]: I0313 09:41:27.099420 4930 scope.go:117] "RemoveContainer" containerID="b831a2f73338ffdf8908b03cae19f580a7caad9da82479cc713725ccc6661253" Mar 13 09:41:34 crc kubenswrapper[4930]: I0313 09:41:34.972030 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:41:34 crc kubenswrapper[4930]: E0313 09:41:34.973042 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:41:38 crc kubenswrapper[4930]: I0313 09:41:38.544134 4930 generic.go:334] "Generic (PLEG): container finished" podID="ab28cafc-95dc-4ace-a257-a428df08e6ef" containerID="0201f8ad55880f3efec8395f9d81bb069ea054bd8cb791509d996e8a00904610" exitCode=0 Mar 13 09:41:38 crc kubenswrapper[4930]: I0313 09:41:38.544209 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"ab28cafc-95dc-4ace-a257-a428df08e6ef","Type":"ContainerDied","Data":"0201f8ad55880f3efec8395f9d81bb069ea054bd8cb791509d996e8a00904610"} Mar 13 09:41:39 crc kubenswrapper[4930]: I0313 09:41:39.559247 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-1" event={"ID":"ab28cafc-95dc-4ace-a257-a428df08e6ef","Type":"ContainerStarted","Data":"b5f96cb804d21df2f8b3b4ab86244543e75e8f78bad30ad4107d1c5474c42f84"} Mar 13 09:41:39 crc kubenswrapper[4930]: I0313 09:41:39.559767 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-1" Mar 13 09:41:39 crc kubenswrapper[4930]: I0313 09:41:39.597769 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-1" podStartSLOduration=41.597749977 podStartE2EDuration="41.597749977s" podCreationTimestamp="2026-03-13 09:40:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:41:39.592126235 +0000 UTC m=+1740.342040912" watchObservedRunningTime="2026-03-13 09:41:39.597749977 +0000 UTC m=+1740.347664654" Mar 13 09:41:47 crc kubenswrapper[4930]: I0313 09:41:47.971453 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:41:47 crc kubenswrapper[4930]: E0313 09:41:47.972473 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:41:49 crc kubenswrapper[4930]: I0313 09:41:49.288744 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-1" Mar 13 09:41:49 crc kubenswrapper[4930]: I0313 09:41:49.370556 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 09:41:53 crc kubenswrapper[4930]: I0313 09:41:53.493971 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="897aa260-40c2-42d2-b59f-964a1a40116c" containerName="rabbitmq" containerID="cri-o://8fa8c46742e2c55c57c3f681b799c6123ebcb666f96bea9223dd69f96d660c17" gracePeriod=604796 Mar 13 09:41:59 crc kubenswrapper[4930]: E0313 09:41:59.770095 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod897aa260_40c2_42d2_b59f_964a1a40116c.slice/crio-8fa8c46742e2c55c57c3f681b799c6123ebcb666f96bea9223dd69f96d660c17.scope\": RecentStats: unable to find data in memory cache]" Mar 13 09:41:59 crc kubenswrapper[4930]: I0313 09:41:59.819025 4930 generic.go:334] "Generic (PLEG): container finished" podID="897aa260-40c2-42d2-b59f-964a1a40116c" containerID="8fa8c46742e2c55c57c3f681b799c6123ebcb666f96bea9223dd69f96d660c17" exitCode=0 Mar 13 09:41:59 crc kubenswrapper[4930]: I0313 09:41:59.819087 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"897aa260-40c2-42d2-b59f-964a1a40116c","Type":"ContainerDied","Data":"8fa8c46742e2c55c57c3f681b799c6123ebcb666f96bea9223dd69f96d660c17"} Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.150418 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556582-g29pq"] Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.152420 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556582-g29pq" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.155344 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.155670 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.155760 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.170386 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556582-g29pq"] Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.260563 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7l4g\" (UniqueName: \"kubernetes.io/projected/c927b774-514d-483a-97ec-d72cbc62a7a3-kube-api-access-x7l4g\") pod \"auto-csr-approver-29556582-g29pq\" (UID: \"c927b774-514d-483a-97ec-d72cbc62a7a3\") " pod="openshift-infra/auto-csr-approver-29556582-g29pq" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.272710 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.363067 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7l4g\" (UniqueName: \"kubernetes.io/projected/c927b774-514d-483a-97ec-d72cbc62a7a3-kube-api-access-x7l4g\") pod \"auto-csr-approver-29556582-g29pq\" (UID: \"c927b774-514d-483a-97ec-d72cbc62a7a3\") " pod="openshift-infra/auto-csr-approver-29556582-g29pq" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.392317 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7l4g\" (UniqueName: \"kubernetes.io/projected/c927b774-514d-483a-97ec-d72cbc62a7a3-kube-api-access-x7l4g\") pod \"auto-csr-approver-29556582-g29pq\" (UID: \"c927b774-514d-483a-97ec-d72cbc62a7a3\") " pod="openshift-infra/auto-csr-approver-29556582-g29pq" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.464923 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-server-conf\") pod \"897aa260-40c2-42d2-b59f-964a1a40116c\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.465005 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-plugins\") pod \"897aa260-40c2-42d2-b59f-964a1a40116c\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.465080 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-confd\") pod \"897aa260-40c2-42d2-b59f-964a1a40116c\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.465153 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-plugins-conf\") pod \"897aa260-40c2-42d2-b59f-964a1a40116c\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.465191 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-config-data\") pod \"897aa260-40c2-42d2-b59f-964a1a40116c\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.465238 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-erlang-cookie\") pod \"897aa260-40c2-42d2-b59f-964a1a40116c\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.465299 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-tls\") pod \"897aa260-40c2-42d2-b59f-964a1a40116c\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.465322 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/897aa260-40c2-42d2-b59f-964a1a40116c-erlang-cookie-secret\") pod \"897aa260-40c2-42d2-b59f-964a1a40116c\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.465373 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jcf7\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-kube-api-access-2jcf7\") pod \"897aa260-40c2-42d2-b59f-964a1a40116c\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.465472 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/897aa260-40c2-42d2-b59f-964a1a40116c-pod-info\") pod \"897aa260-40c2-42d2-b59f-964a1a40116c\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.466038 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\") pod \"897aa260-40c2-42d2-b59f-964a1a40116c\" (UID: \"897aa260-40c2-42d2-b59f-964a1a40116c\") " Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.466292 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "897aa260-40c2-42d2-b59f-964a1a40116c" (UID: "897aa260-40c2-42d2-b59f-964a1a40116c"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.466512 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "897aa260-40c2-42d2-b59f-964a1a40116c" (UID: "897aa260-40c2-42d2-b59f-964a1a40116c"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.467154 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.467181 4930 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.469989 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "897aa260-40c2-42d2-b59f-964a1a40116c" (UID: "897aa260-40c2-42d2-b59f-964a1a40116c"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.470614 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "897aa260-40c2-42d2-b59f-964a1a40116c" (UID: "897aa260-40c2-42d2-b59f-964a1a40116c"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.472198 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-kube-api-access-2jcf7" (OuterVolumeSpecName: "kube-api-access-2jcf7") pod "897aa260-40c2-42d2-b59f-964a1a40116c" (UID: "897aa260-40c2-42d2-b59f-964a1a40116c"). InnerVolumeSpecName "kube-api-access-2jcf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.495848 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/897aa260-40c2-42d2-b59f-964a1a40116c-pod-info" (OuterVolumeSpecName: "pod-info") pod "897aa260-40c2-42d2-b59f-964a1a40116c" (UID: "897aa260-40c2-42d2-b59f-964a1a40116c"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.510627 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/897aa260-40c2-42d2-b59f-964a1a40116c-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "897aa260-40c2-42d2-b59f-964a1a40116c" (UID: "897aa260-40c2-42d2-b59f-964a1a40116c"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.542763 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-config-data" (OuterVolumeSpecName: "config-data") pod "897aa260-40c2-42d2-b59f-964a1a40116c" (UID: "897aa260-40c2-42d2-b59f-964a1a40116c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.544891 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411" (OuterVolumeSpecName: "persistence") pod "897aa260-40c2-42d2-b59f-964a1a40116c" (UID: "897aa260-40c2-42d2-b59f-964a1a40116c"). InnerVolumeSpecName "pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.570403 4930 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\") on node \"crc\" " Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.570462 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.570476 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.570490 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.570501 4930 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/897aa260-40c2-42d2-b59f-964a1a40116c-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.570516 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jcf7\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-kube-api-access-2jcf7\") on node \"crc\" DevicePath \"\"" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.570526 4930 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/897aa260-40c2-42d2-b59f-964a1a40116c-pod-info\") on node \"crc\" DevicePath \"\"" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.583220 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-server-conf" (OuterVolumeSpecName: "server-conf") pod "897aa260-40c2-42d2-b59f-964a1a40116c" (UID: "897aa260-40c2-42d2-b59f-964a1a40116c"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.595052 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556582-g29pq" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.605312 4930 csi_attacher.go:630] kubernetes.io/csi: attacher.UnmountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping UnmountDevice... Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.605496 4930 operation_generator.go:917] UnmountDevice succeeded for volume "pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411") on node "crc" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.669511 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "897aa260-40c2-42d2-b59f-964a1a40116c" (UID: "897aa260-40c2-42d2-b59f-964a1a40116c"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.675817 4930 reconciler_common.go:293] "Volume detached for volume \"pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\") on node \"crc\" DevicePath \"\"" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.675895 4930 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/897aa260-40c2-42d2-b59f-964a1a40116c-server-conf\") on node \"crc\" DevicePath \"\"" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.675918 4930 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/897aa260-40c2-42d2-b59f-964a1a40116c-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.840681 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"897aa260-40c2-42d2-b59f-964a1a40116c","Type":"ContainerDied","Data":"c4bd8903e11193a779d6a36c39193e29bcf85e6394c93049499223a6dcf37350"} Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.840728 4930 scope.go:117] "RemoveContainer" containerID="8fa8c46742e2c55c57c3f681b799c6123ebcb666f96bea9223dd69f96d660c17" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.840869 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.865978 4930 scope.go:117] "RemoveContainer" containerID="5c636643df2852c587885134c9b4420f836bd5a85a714cb57ef6dc4fd70b578a" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.897853 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.955733 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.989866 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 09:42:00 crc kubenswrapper[4930]: E0313 09:42:00.990363 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="897aa260-40c2-42d2-b59f-964a1a40116c" containerName="rabbitmq" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.990388 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="897aa260-40c2-42d2-b59f-964a1a40116c" containerName="rabbitmq" Mar 13 09:42:00 crc kubenswrapper[4930]: E0313 09:42:00.990627 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="897aa260-40c2-42d2-b59f-964a1a40116c" containerName="setup-container" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.990646 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="897aa260-40c2-42d2-b59f-964a1a40116c" containerName="setup-container" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.990983 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="897aa260-40c2-42d2-b59f-964a1a40116c" containerName="rabbitmq" Mar 13 09:42:00 crc kubenswrapper[4930]: I0313 09:42:00.992652 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.005013 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.188887 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b14d5a5a-3717-4c07-9c09-0575088bdee3-config-data\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.189173 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.189262 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b14d5a5a-3717-4c07-9c09-0575088bdee3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.189372 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b14d5a5a-3717-4c07-9c09-0575088bdee3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.189474 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b14d5a5a-3717-4c07-9c09-0575088bdee3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.189578 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b14d5a5a-3717-4c07-9c09-0575088bdee3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.189671 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b14d5a5a-3717-4c07-9c09-0575088bdee3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.194805 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw7ps\" (UniqueName: \"kubernetes.io/projected/b14d5a5a-3717-4c07-9c09-0575088bdee3-kube-api-access-zw7ps\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.195546 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b14d5a5a-3717-4c07-9c09-0575088bdee3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.195714 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b14d5a5a-3717-4c07-9c09-0575088bdee3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.196131 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b14d5a5a-3717-4c07-9c09-0575088bdee3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.196741 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556582-g29pq"] Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.298244 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b14d5a5a-3717-4c07-9c09-0575088bdee3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.298352 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw7ps\" (UniqueName: \"kubernetes.io/projected/b14d5a5a-3717-4c07-9c09-0575088bdee3-kube-api-access-zw7ps\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.298410 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b14d5a5a-3717-4c07-9c09-0575088bdee3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.298536 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b14d5a5a-3717-4c07-9c09-0575088bdee3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.298553 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b14d5a5a-3717-4c07-9c09-0575088bdee3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.298594 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b14d5a5a-3717-4c07-9c09-0575088bdee3-config-data\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.298670 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.298700 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b14d5a5a-3717-4c07-9c09-0575088bdee3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.298752 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b14d5a5a-3717-4c07-9c09-0575088bdee3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.298780 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b14d5a5a-3717-4c07-9c09-0575088bdee3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.298805 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b14d5a5a-3717-4c07-9c09-0575088bdee3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.300249 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/b14d5a5a-3717-4c07-9c09-0575088bdee3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.301205 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/b14d5a5a-3717-4c07-9c09-0575088bdee3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.301544 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/b14d5a5a-3717-4c07-9c09-0575088bdee3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.301736 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/b14d5a5a-3717-4c07-9c09-0575088bdee3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.300870 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/b14d5a5a-3717-4c07-9c09-0575088bdee3-config-data\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.305536 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/b14d5a5a-3717-4c07-9c09-0575088bdee3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.306119 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/b14d5a5a-3717-4c07-9c09-0575088bdee3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.307212 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/b14d5a5a-3717-4c07-9c09-0575088bdee3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.308726 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/b14d5a5a-3717-4c07-9c09-0575088bdee3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.310344 4930 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.310397 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/58fe0acdb728d3cb19cbd87a16786e051248a03f1d289ab57726095b513a2e26/globalmount\"" pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.322048 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw7ps\" (UniqueName: \"kubernetes.io/projected/b14d5a5a-3717-4c07-9c09-0575088bdee3-kube-api-access-zw7ps\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.383046 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-8aac857b-6a55-4ffb-b6e6-e989a6fb8411\") pod \"rabbitmq-server-0\" (UID: \"b14d5a5a-3717-4c07-9c09-0575088bdee3\") " pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.614021 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.855388 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556582-g29pq" event={"ID":"c927b774-514d-483a-97ec-d72cbc62a7a3","Type":"ContainerStarted","Data":"c5c4595d6d91d93d52ca1575d351e5e5dd0d03c068290e37d09d186b81d284e9"} Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.971775 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:42:01 crc kubenswrapper[4930]: E0313 09:42:01.972611 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:42:01 crc kubenswrapper[4930]: I0313 09:42:01.990640 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="897aa260-40c2-42d2-b59f-964a1a40116c" path="/var/lib/kubelet/pods/897aa260-40c2-42d2-b59f-964a1a40116c/volumes" Mar 13 09:42:02 crc kubenswrapper[4930]: I0313 09:42:02.140647 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 13 09:42:02 crc kubenswrapper[4930]: I0313 09:42:02.877827 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556582-g29pq" event={"ID":"c927b774-514d-483a-97ec-d72cbc62a7a3","Type":"ContainerStarted","Data":"3adf722854cc2bf589fb56a6a42c8ea43ad76ee7cc5596cb8c0a20c02739b258"} Mar 13 09:42:02 crc kubenswrapper[4930]: I0313 09:42:02.881912 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b14d5a5a-3717-4c07-9c09-0575088bdee3","Type":"ContainerStarted","Data":"c45ef991ae0e73a69464e54529047e258829cd26656e9b79c16ece3cb54d6bfb"} Mar 13 09:42:02 crc kubenswrapper[4930]: I0313 09:42:02.894343 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556582-g29pq" podStartSLOduration=2.020935888 podStartE2EDuration="2.894320662s" podCreationTimestamp="2026-03-13 09:42:00 +0000 UTC" firstStartedPulling="2026-03-13 09:42:01.199187695 +0000 UTC m=+1761.949102372" lastFinishedPulling="2026-03-13 09:42:02.072572429 +0000 UTC m=+1762.822487146" observedRunningTime="2026-03-13 09:42:02.892565498 +0000 UTC m=+1763.642480195" watchObservedRunningTime="2026-03-13 09:42:02.894320662 +0000 UTC m=+1763.644235359" Mar 13 09:42:03 crc kubenswrapper[4930]: I0313 09:42:03.897232 4930 generic.go:334] "Generic (PLEG): container finished" podID="c927b774-514d-483a-97ec-d72cbc62a7a3" containerID="3adf722854cc2bf589fb56a6a42c8ea43ad76ee7cc5596cb8c0a20c02739b258" exitCode=0 Mar 13 09:42:03 crc kubenswrapper[4930]: I0313 09:42:03.897680 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556582-g29pq" event={"ID":"c927b774-514d-483a-97ec-d72cbc62a7a3","Type":"ContainerDied","Data":"3adf722854cc2bf589fb56a6a42c8ea43ad76ee7cc5596cb8c0a20c02739b258"} Mar 13 09:42:04 crc kubenswrapper[4930]: I0313 09:42:04.920416 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b14d5a5a-3717-4c07-9c09-0575088bdee3","Type":"ContainerStarted","Data":"ad1edf48c93c8377911e5256ce7d290ed073789376ffc1ca027807c4b6c938a7"} Mar 13 09:42:05 crc kubenswrapper[4930]: I0313 09:42:05.353972 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556582-g29pq" Mar 13 09:42:05 crc kubenswrapper[4930]: I0313 09:42:05.440284 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7l4g\" (UniqueName: \"kubernetes.io/projected/c927b774-514d-483a-97ec-d72cbc62a7a3-kube-api-access-x7l4g\") pod \"c927b774-514d-483a-97ec-d72cbc62a7a3\" (UID: \"c927b774-514d-483a-97ec-d72cbc62a7a3\") " Mar 13 09:42:05 crc kubenswrapper[4930]: I0313 09:42:05.445783 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c927b774-514d-483a-97ec-d72cbc62a7a3-kube-api-access-x7l4g" (OuterVolumeSpecName: "kube-api-access-x7l4g") pod "c927b774-514d-483a-97ec-d72cbc62a7a3" (UID: "c927b774-514d-483a-97ec-d72cbc62a7a3"). InnerVolumeSpecName "kube-api-access-x7l4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:42:05 crc kubenswrapper[4930]: I0313 09:42:05.542879 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7l4g\" (UniqueName: \"kubernetes.io/projected/c927b774-514d-483a-97ec-d72cbc62a7a3-kube-api-access-x7l4g\") on node \"crc\" DevicePath \"\"" Mar 13 09:42:05 crc kubenswrapper[4930]: I0313 09:42:05.942037 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556582-g29pq" event={"ID":"c927b774-514d-483a-97ec-d72cbc62a7a3","Type":"ContainerDied","Data":"c5c4595d6d91d93d52ca1575d351e5e5dd0d03c068290e37d09d186b81d284e9"} Mar 13 09:42:05 crc kubenswrapper[4930]: I0313 09:42:05.942398 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c5c4595d6d91d93d52ca1575d351e5e5dd0d03c068290e37d09d186b81d284e9" Mar 13 09:42:05 crc kubenswrapper[4930]: I0313 09:42:05.942068 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556582-g29pq" Mar 13 09:42:05 crc kubenswrapper[4930]: I0313 09:42:05.998402 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556576-kctpg"] Mar 13 09:42:06 crc kubenswrapper[4930]: I0313 09:42:06.026015 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556576-kctpg"] Mar 13 09:42:07 crc kubenswrapper[4930]: I0313 09:42:07.986599 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4ac8095-d474-4fe0-ab66-5e119896d4a1" path="/var/lib/kubelet/pods/f4ac8095-d474-4fe0-ab66-5e119896d4a1/volumes" Mar 13 09:42:14 crc kubenswrapper[4930]: I0313 09:42:14.970822 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:42:14 crc kubenswrapper[4930]: E0313 09:42:14.971639 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:42:27 crc kubenswrapper[4930]: I0313 09:42:27.336357 4930 scope.go:117] "RemoveContainer" containerID="acbe3f948a6eb2d09968373098c52055bb11944d61464e5bbede99904f8b601d" Mar 13 09:42:27 crc kubenswrapper[4930]: I0313 09:42:27.385303 4930 scope.go:117] "RemoveContainer" containerID="07254621feced3ddd66802379857c0ce63823b857833b6250527f980d8badd16" Mar 13 09:42:27 crc kubenswrapper[4930]: I0313 09:42:27.428087 4930 scope.go:117] "RemoveContainer" containerID="74f48ea43202380d568031e44173017842714b290a809a1368e8e7b7a9ab22fa" Mar 13 09:42:27 crc kubenswrapper[4930]: I0313 09:42:27.971045 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:42:27 crc kubenswrapper[4930]: E0313 09:42:27.971484 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:42:36 crc kubenswrapper[4930]: I0313 09:42:36.307456 4930 generic.go:334] "Generic (PLEG): container finished" podID="b14d5a5a-3717-4c07-9c09-0575088bdee3" containerID="ad1edf48c93c8377911e5256ce7d290ed073789376ffc1ca027807c4b6c938a7" exitCode=0 Mar 13 09:42:36 crc kubenswrapper[4930]: I0313 09:42:36.308148 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b14d5a5a-3717-4c07-9c09-0575088bdee3","Type":"ContainerDied","Data":"ad1edf48c93c8377911e5256ce7d290ed073789376ffc1ca027807c4b6c938a7"} Mar 13 09:42:37 crc kubenswrapper[4930]: I0313 09:42:37.324341 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"b14d5a5a-3717-4c07-9c09-0575088bdee3","Type":"ContainerStarted","Data":"4b0a07efa61429779407fb27f9a845f53e3d5973f4bd8e111ecaa76a9f1805d6"} Mar 13 09:42:37 crc kubenswrapper[4930]: I0313 09:42:37.325487 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 13 09:42:37 crc kubenswrapper[4930]: I0313 09:42:37.371568 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.371547786 podStartE2EDuration="37.371547786s" podCreationTimestamp="2026-03-13 09:42:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:42:37.358522975 +0000 UTC m=+1798.108437672" watchObservedRunningTime="2026-03-13 09:42:37.371547786 +0000 UTC m=+1798.121462463" Mar 13 09:42:41 crc kubenswrapper[4930]: I0313 09:42:41.971961 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:42:41 crc kubenswrapper[4930]: E0313 09:42:41.974156 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:42:51 crc kubenswrapper[4930]: I0313 09:42:51.618575 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 13 09:42:53 crc kubenswrapper[4930]: I0313 09:42:53.971456 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:42:53 crc kubenswrapper[4930]: E0313 09:42:53.972157 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:43:06 crc kubenswrapper[4930]: I0313 09:43:06.971381 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:43:06 crc kubenswrapper[4930]: E0313 09:43:06.972134 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:43:21 crc kubenswrapper[4930]: I0313 09:43:21.975689 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:43:21 crc kubenswrapper[4930]: E0313 09:43:21.976595 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:43:27 crc kubenswrapper[4930]: I0313 09:43:27.544260 4930 scope.go:117] "RemoveContainer" containerID="293f4ccae5256f85e7a553d009bf9d4a4ec9b9385f9b2557386cced5fe7d45ab" Mar 13 09:43:32 crc kubenswrapper[4930]: I0313 09:43:32.971044 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:43:32 crc kubenswrapper[4930]: E0313 09:43:32.971911 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:43:43 crc kubenswrapper[4930]: I0313 09:43:43.083621 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-hvlz5"] Mar 13 09:43:43 crc kubenswrapper[4930]: I0313 09:43:43.110324 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-5f25-account-create-update-t9v5n"] Mar 13 09:43:43 crc kubenswrapper[4930]: I0313 09:43:43.133284 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-0a6a-account-create-update-kvkqr"] Mar 13 09:43:43 crc kubenswrapper[4930]: I0313 09:43:43.149391 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-k6ktf"] Mar 13 09:43:43 crc kubenswrapper[4930]: I0313 09:43:43.161694 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-hvlz5"] Mar 13 09:43:43 crc kubenswrapper[4930]: I0313 09:43:43.171725 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-0a6a-account-create-update-kvkqr"] Mar 13 09:43:43 crc kubenswrapper[4930]: I0313 09:43:43.183744 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-5f25-account-create-update-t9v5n"] Mar 13 09:43:43 crc kubenswrapper[4930]: I0313 09:43:43.196259 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-k6ktf"] Mar 13 09:43:43 crc kubenswrapper[4930]: I0313 09:43:43.985835 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1731159-7f86-4cbe-b9cb-53e28cc38092" path="/var/lib/kubelet/pods/c1731159-7f86-4cbe-b9cb-53e28cc38092/volumes" Mar 13 09:43:43 crc kubenswrapper[4930]: I0313 09:43:43.991597 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc952175-d111-4e8c-972d-e26cca11ba49" path="/var/lib/kubelet/pods/cc952175-d111-4e8c-972d-e26cca11ba49/volumes" Mar 13 09:43:43 crc kubenswrapper[4930]: I0313 09:43:43.995892 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed73f0c1-2715-409f-9a4a-9da4fa8a3f48" path="/var/lib/kubelet/pods/ed73f0c1-2715-409f-9a4a-9da4fa8a3f48/volumes" Mar 13 09:43:43 crc kubenswrapper[4930]: I0313 09:43:43.998297 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eebc86bf-e7ee-4465-bd15-bef833ac0843" path="/var/lib/kubelet/pods/eebc86bf-e7ee-4465-bd15-bef833ac0843/volumes" Mar 13 09:43:44 crc kubenswrapper[4930]: I0313 09:43:44.970591 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:43:45 crc kubenswrapper[4930]: I0313 09:43:45.555985 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"9d9e501a7dfec9be2a4a7aae030dcc1a1a33117fc6ee6f36debdcb82b8cfee89"} Mar 13 09:43:48 crc kubenswrapper[4930]: I0313 09:43:48.036730 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-ptkns"] Mar 13 09:43:48 crc kubenswrapper[4930]: I0313 09:43:48.053284 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-ptkns"] Mar 13 09:43:48 crc kubenswrapper[4930]: I0313 09:43:48.071138 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-bqwkt"] Mar 13 09:43:48 crc kubenswrapper[4930]: I0313 09:43:48.089223 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-5fae-account-create-update-tdfsn"] Mar 13 09:43:48 crc kubenswrapper[4930]: I0313 09:43:48.100403 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-8baa-account-create-update-b5w9f"] Mar 13 09:43:48 crc kubenswrapper[4930]: I0313 09:43:48.126910 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-bqwkt"] Mar 13 09:43:48 crc kubenswrapper[4930]: I0313 09:43:48.138535 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-5fae-account-create-update-tdfsn"] Mar 13 09:43:48 crc kubenswrapper[4930]: I0313 09:43:48.155899 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-8baa-account-create-update-b5w9f"] Mar 13 09:43:49 crc kubenswrapper[4930]: I0313 09:43:49.031603 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-44qz9"] Mar 13 09:43:49 crc kubenswrapper[4930]: I0313 09:43:49.049126 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-44qz9"] Mar 13 09:43:50 crc kubenswrapper[4930]: I0313 09:43:50.004089 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0d7d0626-3583-4461-957d-d978d8f78ed4" path="/var/lib/kubelet/pods/0d7d0626-3583-4461-957d-d978d8f78ed4/volumes" Mar 13 09:43:50 crc kubenswrapper[4930]: I0313 09:43:50.006853 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="451992e0-ca47-4af8-8b22-54e3c955b3aa" path="/var/lib/kubelet/pods/451992e0-ca47-4af8-8b22-54e3c955b3aa/volumes" Mar 13 09:43:50 crc kubenswrapper[4930]: I0313 09:43:50.007988 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="902fb09a-cf93-4d3c-95d8-6098cb86b34b" path="/var/lib/kubelet/pods/902fb09a-cf93-4d3c-95d8-6098cb86b34b/volumes" Mar 13 09:43:50 crc kubenswrapper[4930]: I0313 09:43:50.009504 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b0fe8888-5e4f-4bec-a5bf-554c209325ed" path="/var/lib/kubelet/pods/b0fe8888-5e4f-4bec-a5bf-554c209325ed/volumes" Mar 13 09:43:50 crc kubenswrapper[4930]: I0313 09:43:50.010603 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3efe40c-96ae-4b9e-a787-7781db5af486" path="/var/lib/kubelet/pods/f3efe40c-96ae-4b9e-a787-7781db5af486/volumes" Mar 13 09:43:51 crc kubenswrapper[4930]: I0313 09:43:51.622598 4930 generic.go:334] "Generic (PLEG): container finished" podID="1c483489-be42-4344-8485-7959520070cb" containerID="fa64ce7da113407390ca4c267e066263118e71079a856424835f34cb916df480" exitCode=0 Mar 13 09:43:51 crc kubenswrapper[4930]: I0313 09:43:51.622696 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" event={"ID":"1c483489-be42-4344-8485-7959520070cb","Type":"ContainerDied","Data":"fa64ce7da113407390ca4c267e066263118e71079a856424835f34cb916df480"} Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.165353 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.358708 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-ssh-key-openstack-edpm-ipam\") pod \"1c483489-be42-4344-8485-7959520070cb\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.359179 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-inventory\") pod \"1c483489-be42-4344-8485-7959520070cb\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.359235 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-bootstrap-combined-ca-bundle\") pod \"1c483489-be42-4344-8485-7959520070cb\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.359403 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8rkt\" (UniqueName: \"kubernetes.io/projected/1c483489-be42-4344-8485-7959520070cb-kube-api-access-j8rkt\") pod \"1c483489-be42-4344-8485-7959520070cb\" (UID: \"1c483489-be42-4344-8485-7959520070cb\") " Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.365999 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "1c483489-be42-4344-8485-7959520070cb" (UID: "1c483489-be42-4344-8485-7959520070cb"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.370568 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c483489-be42-4344-8485-7959520070cb-kube-api-access-j8rkt" (OuterVolumeSpecName: "kube-api-access-j8rkt") pod "1c483489-be42-4344-8485-7959520070cb" (UID: "1c483489-be42-4344-8485-7959520070cb"). InnerVolumeSpecName "kube-api-access-j8rkt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.396005 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-inventory" (OuterVolumeSpecName: "inventory") pod "1c483489-be42-4344-8485-7959520070cb" (UID: "1c483489-be42-4344-8485-7959520070cb"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.396551 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1c483489-be42-4344-8485-7959520070cb" (UID: "1c483489-be42-4344-8485-7959520070cb"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.462621 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.462698 4930 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.462714 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8rkt\" (UniqueName: \"kubernetes.io/projected/1c483489-be42-4344-8485-7959520070cb-kube-api-access-j8rkt\") on node \"crc\" DevicePath \"\"" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.462728 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1c483489-be42-4344-8485-7959520070cb-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.647995 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" event={"ID":"1c483489-be42-4344-8485-7959520070cb","Type":"ContainerDied","Data":"37d45f544ef16806c6978c45044592f8951c838b0651b98d7e66103fb7a2e920"} Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.648040 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-46x4s" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.648047 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="37d45f544ef16806c6978c45044592f8951c838b0651b98d7e66103fb7a2e920" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.752493 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m"] Mar 13 09:43:53 crc kubenswrapper[4930]: E0313 09:43:53.753233 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c927b774-514d-483a-97ec-d72cbc62a7a3" containerName="oc" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.753312 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c927b774-514d-483a-97ec-d72cbc62a7a3" containerName="oc" Mar 13 09:43:53 crc kubenswrapper[4930]: E0313 09:43:53.753402 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c483489-be42-4344-8485-7959520070cb" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.753469 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c483489-be42-4344-8485-7959520070cb" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.753730 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c927b774-514d-483a-97ec-d72cbc62a7a3" containerName="oc" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.753809 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c483489-be42-4344-8485-7959520070cb" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.754668 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.759467 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.759694 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.759815 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.760330 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.773394 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3331ecec-adf5-4107-a891-07d2fd6e4be7-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m\" (UID: \"3331ecec-adf5-4107-a891-07d2fd6e4be7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.773811 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3331ecec-adf5-4107-a891-07d2fd6e4be7-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m\" (UID: \"3331ecec-adf5-4107-a891-07d2fd6e4be7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.773993 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpf27\" (UniqueName: \"kubernetes.io/projected/3331ecec-adf5-4107-a891-07d2fd6e4be7-kube-api-access-cpf27\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m\" (UID: \"3331ecec-adf5-4107-a891-07d2fd6e4be7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.775667 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m"] Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.881068 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3331ecec-adf5-4107-a891-07d2fd6e4be7-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m\" (UID: \"3331ecec-adf5-4107-a891-07d2fd6e4be7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.881193 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3331ecec-adf5-4107-a891-07d2fd6e4be7-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m\" (UID: \"3331ecec-adf5-4107-a891-07d2fd6e4be7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.881295 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpf27\" (UniqueName: \"kubernetes.io/projected/3331ecec-adf5-4107-a891-07d2fd6e4be7-kube-api-access-cpf27\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m\" (UID: \"3331ecec-adf5-4107-a891-07d2fd6e4be7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.885281 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3331ecec-adf5-4107-a891-07d2fd6e4be7-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m\" (UID: \"3331ecec-adf5-4107-a891-07d2fd6e4be7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.885943 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3331ecec-adf5-4107-a891-07d2fd6e4be7-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m\" (UID: \"3331ecec-adf5-4107-a891-07d2fd6e4be7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:43:53 crc kubenswrapper[4930]: I0313 09:43:53.911107 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpf27\" (UniqueName: \"kubernetes.io/projected/3331ecec-adf5-4107-a891-07d2fd6e4be7-kube-api-access-cpf27\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m\" (UID: \"3331ecec-adf5-4107-a891-07d2fd6e4be7\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:43:54 crc kubenswrapper[4930]: I0313 09:43:54.078104 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:43:54 crc kubenswrapper[4930]: I0313 09:43:54.621847 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m"] Mar 13 09:43:54 crc kubenswrapper[4930]: I0313 09:43:54.669031 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" event={"ID":"3331ecec-adf5-4107-a891-07d2fd6e4be7","Type":"ContainerStarted","Data":"5c759cf6072d6eae031caeffbf59585ab2a9319e53341657dfd8bdf739870246"} Mar 13 09:43:55 crc kubenswrapper[4930]: I0313 09:43:55.680931 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" event={"ID":"3331ecec-adf5-4107-a891-07d2fd6e4be7","Type":"ContainerStarted","Data":"fc8b94642ed37d8e6adc9b63f377e6cee6aea018e3f794e0ea1d2f64fc82ea58"} Mar 13 09:43:55 crc kubenswrapper[4930]: I0313 09:43:55.698802 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" podStartSLOduration=1.987788563 podStartE2EDuration="2.698781483s" podCreationTimestamp="2026-03-13 09:43:53 +0000 UTC" firstStartedPulling="2026-03-13 09:43:54.625649989 +0000 UTC m=+1875.375564666" lastFinishedPulling="2026-03-13 09:43:55.336642919 +0000 UTC m=+1876.086557586" observedRunningTime="2026-03-13 09:43:55.697837429 +0000 UTC m=+1876.447752106" watchObservedRunningTime="2026-03-13 09:43:55.698781483 +0000 UTC m=+1876.448696160" Mar 13 09:43:58 crc kubenswrapper[4930]: I0313 09:43:58.052338 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz"] Mar 13 09:43:58 crc kubenswrapper[4930]: I0313 09:43:58.070338 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-b8gsz"] Mar 13 09:43:58 crc kubenswrapper[4930]: I0313 09:43:58.083179 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0d69-account-create-update-rg6d8"] Mar 13 09:43:58 crc kubenswrapper[4930]: I0313 09:43:58.097443 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0d69-account-create-update-rg6d8"] Mar 13 09:43:59 crc kubenswrapper[4930]: I0313 09:43:59.984960 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c20e5a2-621a-413c-8109-121f96f217b8" path="/var/lib/kubelet/pods/4c20e5a2-621a-413c-8109-121f96f217b8/volumes" Mar 13 09:43:59 crc kubenswrapper[4930]: I0313 09:43:59.986094 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="657e3e90-4b08-4df2-b773-1bbc1add41a7" path="/var/lib/kubelet/pods/657e3e90-4b08-4df2-b773-1bbc1add41a7/volumes" Mar 13 09:44:00 crc kubenswrapper[4930]: I0313 09:44:00.133343 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556584-qmpnl"] Mar 13 09:44:00 crc kubenswrapper[4930]: I0313 09:44:00.135004 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556584-qmpnl" Mar 13 09:44:00 crc kubenswrapper[4930]: I0313 09:44:00.137399 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:44:00 crc kubenswrapper[4930]: I0313 09:44:00.137789 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:44:00 crc kubenswrapper[4930]: I0313 09:44:00.137930 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:44:00 crc kubenswrapper[4930]: I0313 09:44:00.145738 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556584-qmpnl"] Mar 13 09:44:00 crc kubenswrapper[4930]: I0313 09:44:00.241236 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ff2td\" (UniqueName: \"kubernetes.io/projected/ef9b7b54-f041-4a1f-84bf-f8b2ee52818d-kube-api-access-ff2td\") pod \"auto-csr-approver-29556584-qmpnl\" (UID: \"ef9b7b54-f041-4a1f-84bf-f8b2ee52818d\") " pod="openshift-infra/auto-csr-approver-29556584-qmpnl" Mar 13 09:44:00 crc kubenswrapper[4930]: I0313 09:44:00.344156 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ff2td\" (UniqueName: \"kubernetes.io/projected/ef9b7b54-f041-4a1f-84bf-f8b2ee52818d-kube-api-access-ff2td\") pod \"auto-csr-approver-29556584-qmpnl\" (UID: \"ef9b7b54-f041-4a1f-84bf-f8b2ee52818d\") " pod="openshift-infra/auto-csr-approver-29556584-qmpnl" Mar 13 09:44:00 crc kubenswrapper[4930]: I0313 09:44:00.371722 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ff2td\" (UniqueName: \"kubernetes.io/projected/ef9b7b54-f041-4a1f-84bf-f8b2ee52818d-kube-api-access-ff2td\") pod \"auto-csr-approver-29556584-qmpnl\" (UID: \"ef9b7b54-f041-4a1f-84bf-f8b2ee52818d\") " pod="openshift-infra/auto-csr-approver-29556584-qmpnl" Mar 13 09:44:00 crc kubenswrapper[4930]: I0313 09:44:00.456897 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556584-qmpnl" Mar 13 09:44:01 crc kubenswrapper[4930]: I0313 09:44:01.027189 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556584-qmpnl"] Mar 13 09:44:01 crc kubenswrapper[4930]: I0313 09:44:01.753499 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556584-qmpnl" event={"ID":"ef9b7b54-f041-4a1f-84bf-f8b2ee52818d","Type":"ContainerStarted","Data":"e1c7b330659ef44cd1d8fe0cf8767920078b6febb7db371c565c6e1c6f31f4c1"} Mar 13 09:44:02 crc kubenswrapper[4930]: I0313 09:44:02.767643 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556584-qmpnl" event={"ID":"ef9b7b54-f041-4a1f-84bf-f8b2ee52818d","Type":"ContainerStarted","Data":"7fcf530a30accc88731658ebc0a7c9693a939ead6904aeeaeefeda5b176ec3dc"} Mar 13 09:44:02 crc kubenswrapper[4930]: I0313 09:44:02.806210 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556584-qmpnl" podStartSLOduration=1.5796093199999999 podStartE2EDuration="2.806181975s" podCreationTimestamp="2026-03-13 09:44:00 +0000 UTC" firstStartedPulling="2026-03-13 09:44:01.058697093 +0000 UTC m=+1881.808611770" lastFinishedPulling="2026-03-13 09:44:02.285269748 +0000 UTC m=+1883.035184425" observedRunningTime="2026-03-13 09:44:02.786414256 +0000 UTC m=+1883.536328943" watchObservedRunningTime="2026-03-13 09:44:02.806181975 +0000 UTC m=+1883.556096652" Mar 13 09:44:03 crc kubenswrapper[4930]: I0313 09:44:03.782072 4930 generic.go:334] "Generic (PLEG): container finished" podID="ef9b7b54-f041-4a1f-84bf-f8b2ee52818d" containerID="7fcf530a30accc88731658ebc0a7c9693a939ead6904aeeaeefeda5b176ec3dc" exitCode=0 Mar 13 09:44:03 crc kubenswrapper[4930]: I0313 09:44:03.782128 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556584-qmpnl" event={"ID":"ef9b7b54-f041-4a1f-84bf-f8b2ee52818d","Type":"ContainerDied","Data":"7fcf530a30accc88731658ebc0a7c9693a939ead6904aeeaeefeda5b176ec3dc"} Mar 13 09:44:05 crc kubenswrapper[4930]: I0313 09:44:05.208656 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556584-qmpnl" Mar 13 09:44:05 crc kubenswrapper[4930]: I0313 09:44:05.280383 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ff2td\" (UniqueName: \"kubernetes.io/projected/ef9b7b54-f041-4a1f-84bf-f8b2ee52818d-kube-api-access-ff2td\") pod \"ef9b7b54-f041-4a1f-84bf-f8b2ee52818d\" (UID: \"ef9b7b54-f041-4a1f-84bf-f8b2ee52818d\") " Mar 13 09:44:05 crc kubenswrapper[4930]: I0313 09:44:05.289261 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ef9b7b54-f041-4a1f-84bf-f8b2ee52818d-kube-api-access-ff2td" (OuterVolumeSpecName: "kube-api-access-ff2td") pod "ef9b7b54-f041-4a1f-84bf-f8b2ee52818d" (UID: "ef9b7b54-f041-4a1f-84bf-f8b2ee52818d"). InnerVolumeSpecName "kube-api-access-ff2td". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:44:05 crc kubenswrapper[4930]: I0313 09:44:05.382204 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ff2td\" (UniqueName: \"kubernetes.io/projected/ef9b7b54-f041-4a1f-84bf-f8b2ee52818d-kube-api-access-ff2td\") on node \"crc\" DevicePath \"\"" Mar 13 09:44:05 crc kubenswrapper[4930]: I0313 09:44:05.806816 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556584-qmpnl" event={"ID":"ef9b7b54-f041-4a1f-84bf-f8b2ee52818d","Type":"ContainerDied","Data":"e1c7b330659ef44cd1d8fe0cf8767920078b6febb7db371c565c6e1c6f31f4c1"} Mar 13 09:44:05 crc kubenswrapper[4930]: I0313 09:44:05.807157 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e1c7b330659ef44cd1d8fe0cf8767920078b6febb7db371c565c6e1c6f31f4c1" Mar 13 09:44:05 crc kubenswrapper[4930]: I0313 09:44:05.806874 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556584-qmpnl" Mar 13 09:44:05 crc kubenswrapper[4930]: I0313 09:44:05.866049 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556578-qw9pg"] Mar 13 09:44:05 crc kubenswrapper[4930]: I0313 09:44:05.876507 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556578-qw9pg"] Mar 13 09:44:05 crc kubenswrapper[4930]: I0313 09:44:05.986310 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1645d8f-1162-44a6-a9df-9bb32862d73c" path="/var/lib/kubelet/pods/a1645d8f-1162-44a6-a9df-9bb32862d73c/volumes" Mar 13 09:44:27 crc kubenswrapper[4930]: I0313 09:44:27.611705 4930 scope.go:117] "RemoveContainer" containerID="50c93bba51bfa60c836f61ac84f93bcb53ca19ac740d2b512126ac639de9b8bf" Mar 13 09:44:27 crc kubenswrapper[4930]: I0313 09:44:27.673173 4930 scope.go:117] "RemoveContainer" containerID="c0c8404cd93a43566c8682cea787be016fc13cfa6d6f53223169bbcd76373fb2" Mar 13 09:44:27 crc kubenswrapper[4930]: I0313 09:44:27.727388 4930 scope.go:117] "RemoveContainer" containerID="c45ac4a97f6a89041787ac9d397f7972f50c469a10f0e15b010101bf8dc2831b" Mar 13 09:44:27 crc kubenswrapper[4930]: I0313 09:44:27.769251 4930 scope.go:117] "RemoveContainer" containerID="01948896c7e83178c616925089deed8a34a3b95d918a9076a1e2364e6e254ddb" Mar 13 09:44:27 crc kubenswrapper[4930]: I0313 09:44:27.815576 4930 scope.go:117] "RemoveContainer" containerID="fa5304a13244c0dad9be995b80117f885c94a201f4733fc13ee62cc0a378d27b" Mar 13 09:44:27 crc kubenswrapper[4930]: I0313 09:44:27.870265 4930 scope.go:117] "RemoveContainer" containerID="08020393b80436228d130190c9583a49ddda5ae786b94588b37ff58a1ad73db6" Mar 13 09:44:27 crc kubenswrapper[4930]: I0313 09:44:27.929218 4930 scope.go:117] "RemoveContainer" containerID="2ee5212ad57c16330e57fe3aed5020f51bdd1d7af3441124b4fbc62930021d66" Mar 13 09:44:27 crc kubenswrapper[4930]: I0313 09:44:27.956772 4930 scope.go:117] "RemoveContainer" containerID="10b2cbb83b2d4a9450d5630b7c7bd2ee7ffb3c4f9cfb18964c42792c439ccf1a" Mar 13 09:44:28 crc kubenswrapper[4930]: I0313 09:44:28.017298 4930 scope.go:117] "RemoveContainer" containerID="e37e8c0afc0bbe36080c34a4600320ec76139a7878514ab22d546557609d3afc" Mar 13 09:44:28 crc kubenswrapper[4930]: I0313 09:44:28.038412 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-ddz9m"] Mar 13 09:44:28 crc kubenswrapper[4930]: I0313 09:44:28.049469 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-ddz9m"] Mar 13 09:44:28 crc kubenswrapper[4930]: I0313 09:44:28.070892 4930 scope.go:117] "RemoveContainer" containerID="539a0f8a7b2944548c0bc24d36814e3b7804441ec50a4b08b916c45d1cde6c67" Mar 13 09:44:28 crc kubenswrapper[4930]: I0313 09:44:28.124259 4930 scope.go:117] "RemoveContainer" containerID="44c06115d18097c2cb7c6cfe31705ac6acb7f5865b83b24e0aeddcb7ffa980cf" Mar 13 09:44:28 crc kubenswrapper[4930]: I0313 09:44:28.146381 4930 scope.go:117] "RemoveContainer" containerID="0fd7a851ab276bfdb2cef98055c533b02ef0439cf4fc96152c66fafa107369e8" Mar 13 09:44:28 crc kubenswrapper[4930]: I0313 09:44:28.175866 4930 scope.go:117] "RemoveContainer" containerID="8f1a17ded6e885d9cba84dfb6f06c4cce0223ed0c64ea7df78c1c8c99a8e3611" Mar 13 09:44:30 crc kubenswrapper[4930]: I0313 09:44:30.010182 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86a246ca-6999-4a55-aa0b-b723006d6305" path="/var/lib/kubelet/pods/86a246ca-6999-4a55-aa0b-b723006d6305/volumes" Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.040246 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-bdfd-account-create-update-9lgz8"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.061940 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-bdfd-account-create-update-9lgz8"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.075683 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-0b34-account-create-update-grx5t"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.091512 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-0b34-account-create-update-grx5t"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.104098 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-pbbn5"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.115007 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-zqbhp"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.126507 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-728c-account-create-update-q5spq"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.136683 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-lgr8r"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.146338 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-728c-account-create-update-q5spq"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.157349 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-pbbn5"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.168808 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-zqbhp"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.181342 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-lgr8r"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.192525 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-d7cf-account-create-update-v5gw9"] Mar 13 09:44:32 crc kubenswrapper[4930]: I0313 09:44:32.202905 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-d7cf-account-create-update-v5gw9"] Mar 13 09:44:33 crc kubenswrapper[4930]: I0313 09:44:33.984429 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6164de25-5015-4f67-8b21-d0203eae1351" path="/var/lib/kubelet/pods/6164de25-5015-4f67-8b21-d0203eae1351/volumes" Mar 13 09:44:33 crc kubenswrapper[4930]: I0313 09:44:33.987745 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f31cf24-9bb2-41c6-a825-c5481a27ffb0" path="/var/lib/kubelet/pods/8f31cf24-9bb2-41c6-a825-c5481a27ffb0/volumes" Mar 13 09:44:33 crc kubenswrapper[4930]: I0313 09:44:33.988511 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95e1aca8-117b-43da-b0ce-3b132c4a6c60" path="/var/lib/kubelet/pods/95e1aca8-117b-43da-b0ce-3b132c4a6c60/volumes" Mar 13 09:44:33 crc kubenswrapper[4930]: I0313 09:44:33.990048 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ade20871-d4bc-4a8f-9635-d28bbae44421" path="/var/lib/kubelet/pods/ade20871-d4bc-4a8f-9635-d28bbae44421/volumes" Mar 13 09:44:33 crc kubenswrapper[4930]: I0313 09:44:33.990694 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caf815ba-1554-4c46-bc4d-38b8106cced5" path="/var/lib/kubelet/pods/caf815ba-1554-4c46-bc4d-38b8106cced5/volumes" Mar 13 09:44:33 crc kubenswrapper[4930]: I0313 09:44:33.991775 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f84311c2-0fb2-4d15-8272-f245fcbc9957" path="/var/lib/kubelet/pods/f84311c2-0fb2-4d15-8272-f245fcbc9957/volumes" Mar 13 09:44:33 crc kubenswrapper[4930]: I0313 09:44:33.992368 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fffd6b7f-d18d-4ff8-a446-032de1c3e288" path="/var/lib/kubelet/pods/fffd6b7f-d18d-4ff8-a446-032de1c3e288/volumes" Mar 13 09:44:37 crc kubenswrapper[4930]: I0313 09:44:37.051199 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-vr4gs"] Mar 13 09:44:37 crc kubenswrapper[4930]: I0313 09:44:37.066290 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-vr4gs"] Mar 13 09:44:37 crc kubenswrapper[4930]: I0313 09:44:37.983951 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0127857-fa77-4179-ad95-7dfb2585169a" path="/var/lib/kubelet/pods/d0127857-fa77-4179-ad95-7dfb2585169a/volumes" Mar 13 09:44:41 crc kubenswrapper[4930]: I0313 09:44:41.033774 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-g82c9"] Mar 13 09:44:41 crc kubenswrapper[4930]: I0313 09:44:41.045096 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-g82c9"] Mar 13 09:44:41 crc kubenswrapper[4930]: I0313 09:44:41.986336 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf10d12f-578b-400e-b480-60a7343bc344" path="/var/lib/kubelet/pods/bf10d12f-578b-400e-b480-60a7343bc344/volumes" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.156693 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb"] Mar 13 09:45:00 crc kubenswrapper[4930]: E0313 09:45:00.157758 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ef9b7b54-f041-4a1f-84bf-f8b2ee52818d" containerName="oc" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.157964 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ef9b7b54-f041-4a1f-84bf-f8b2ee52818d" containerName="oc" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.158217 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ef9b7b54-f041-4a1f-84bf-f8b2ee52818d" containerName="oc" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.159115 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.162152 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.162135 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.173069 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb"] Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.297469 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-config-volume\") pod \"collect-profiles-29556585-5qpzb\" (UID: \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.298220 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8r25k\" (UniqueName: \"kubernetes.io/projected/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-kube-api-access-8r25k\") pod \"collect-profiles-29556585-5qpzb\" (UID: \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.298580 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-secret-volume\") pod \"collect-profiles-29556585-5qpzb\" (UID: \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.401322 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8r25k\" (UniqueName: \"kubernetes.io/projected/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-kube-api-access-8r25k\") pod \"collect-profiles-29556585-5qpzb\" (UID: \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.401444 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-secret-volume\") pod \"collect-profiles-29556585-5qpzb\" (UID: \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.401493 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-config-volume\") pod \"collect-profiles-29556585-5qpzb\" (UID: \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.402749 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-config-volume\") pod \"collect-profiles-29556585-5qpzb\" (UID: \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.407310 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-secret-volume\") pod \"collect-profiles-29556585-5qpzb\" (UID: \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.418116 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8r25k\" (UniqueName: \"kubernetes.io/projected/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-kube-api-access-8r25k\") pod \"collect-profiles-29556585-5qpzb\" (UID: \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.485531 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:00 crc kubenswrapper[4930]: I0313 09:45:00.974335 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb"] Mar 13 09:45:01 crc kubenswrapper[4930]: I0313 09:45:01.446071 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" event={"ID":"ed8cfe62-57c8-42c3-82bb-d712ff3f8265","Type":"ContainerStarted","Data":"d55e708d8732d2a3a27eb859cbb1cbb1e26816bbf9ee88fb2731459c1b608416"} Mar 13 09:45:01 crc kubenswrapper[4930]: I0313 09:45:01.446118 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" event={"ID":"ed8cfe62-57c8-42c3-82bb-d712ff3f8265","Type":"ContainerStarted","Data":"c572957cfa07eb2551ee8a7d47b1a00b8139c3696972622739691be5a488554f"} Mar 13 09:45:01 crc kubenswrapper[4930]: I0313 09:45:01.472472 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" podStartSLOduration=1.472454243 podStartE2EDuration="1.472454243s" podCreationTimestamp="2026-03-13 09:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 09:45:01.463314022 +0000 UTC m=+1942.213228699" watchObservedRunningTime="2026-03-13 09:45:01.472454243 +0000 UTC m=+1942.222368920" Mar 13 09:45:02 crc kubenswrapper[4930]: I0313 09:45:02.470894 4930 generic.go:334] "Generic (PLEG): container finished" podID="ed8cfe62-57c8-42c3-82bb-d712ff3f8265" containerID="d55e708d8732d2a3a27eb859cbb1cbb1e26816bbf9ee88fb2731459c1b608416" exitCode=0 Mar 13 09:45:02 crc kubenswrapper[4930]: I0313 09:45:02.471364 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" event={"ID":"ed8cfe62-57c8-42c3-82bb-d712ff3f8265","Type":"ContainerDied","Data":"d55e708d8732d2a3a27eb859cbb1cbb1e26816bbf9ee88fb2731459c1b608416"} Mar 13 09:45:03 crc kubenswrapper[4930]: I0313 09:45:03.923327 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.102308 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-secret-volume\") pod \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\" (UID: \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\") " Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.102475 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8r25k\" (UniqueName: \"kubernetes.io/projected/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-kube-api-access-8r25k\") pod \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\" (UID: \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\") " Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.102638 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-config-volume\") pod \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\" (UID: \"ed8cfe62-57c8-42c3-82bb-d712ff3f8265\") " Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.103646 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-config-volume" (OuterVolumeSpecName: "config-volume") pod "ed8cfe62-57c8-42c3-82bb-d712ff3f8265" (UID: "ed8cfe62-57c8-42c3-82bb-d712ff3f8265"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.108413 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-kube-api-access-8r25k" (OuterVolumeSpecName: "kube-api-access-8r25k") pod "ed8cfe62-57c8-42c3-82bb-d712ff3f8265" (UID: "ed8cfe62-57c8-42c3-82bb-d712ff3f8265"). InnerVolumeSpecName "kube-api-access-8r25k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.108747 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "ed8cfe62-57c8-42c3-82bb-d712ff3f8265" (UID: "ed8cfe62-57c8-42c3-82bb-d712ff3f8265"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.205966 4930 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.206470 4930 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.206503 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8r25k\" (UniqueName: \"kubernetes.io/projected/ed8cfe62-57c8-42c3-82bb-d712ff3f8265-kube-api-access-8r25k\") on node \"crc\" DevicePath \"\"" Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.499152 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" event={"ID":"ed8cfe62-57c8-42c3-82bb-d712ff3f8265","Type":"ContainerDied","Data":"c572957cfa07eb2551ee8a7d47b1a00b8139c3696972622739691be5a488554f"} Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.499196 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c572957cfa07eb2551ee8a7d47b1a00b8139c3696972622739691be5a488554f" Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.499273 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb" Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.548831 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz"] Mar 13 09:45:04 crc kubenswrapper[4930]: I0313 09:45:04.563633 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556540-w2lkz"] Mar 13 09:45:06 crc kubenswrapper[4930]: I0313 09:45:06.029861 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f65639e5-a600-44c3-a8b5-49aea65404e0" path="/var/lib/kubelet/pods/f65639e5-a600-44c3-a8b5-49aea65404e0/volumes" Mar 13 09:45:21 crc kubenswrapper[4930]: I0313 09:45:21.060591 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-pdp9j"] Mar 13 09:45:21 crc kubenswrapper[4930]: I0313 09:45:21.073279 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-pdp9j"] Mar 13 09:45:21 crc kubenswrapper[4930]: I0313 09:45:21.992730 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d37e8e55-e1f6-4c48-8f14-e4d714afb1fc" path="/var/lib/kubelet/pods/d37e8e55-e1f6-4c48-8f14-e4d714afb1fc/volumes" Mar 13 09:45:28 crc kubenswrapper[4930]: I0313 09:45:28.528250 4930 scope.go:117] "RemoveContainer" containerID="9cd8e9f98e28ab2aa6c6fa08257cd36d1c89eaf1805033449f57940002c31365" Mar 13 09:45:28 crc kubenswrapper[4930]: I0313 09:45:28.559723 4930 scope.go:117] "RemoveContainer" containerID="782485b9d9f78d2c982187e6bb785310dfbbf0b35c611d44d7585c989975a169" Mar 13 09:45:28 crc kubenswrapper[4930]: I0313 09:45:28.626892 4930 scope.go:117] "RemoveContainer" containerID="2d261f4b77a94477c1f9a893bd96a0a2e94a766d584dc966db7302f165125734" Mar 13 09:45:28 crc kubenswrapper[4930]: I0313 09:45:28.678826 4930 scope.go:117] "RemoveContainer" containerID="2d49098b9512d6183bdf64a950a657e8445ed19536c9fc23675722560104044b" Mar 13 09:45:28 crc kubenswrapper[4930]: I0313 09:45:28.733699 4930 scope.go:117] "RemoveContainer" containerID="21eed8e0619bb514a23b49eb0bab1cf671d8a7d01d4d46d3b1b3946831c8583f" Mar 13 09:45:28 crc kubenswrapper[4930]: I0313 09:45:28.791828 4930 scope.go:117] "RemoveContainer" containerID="189d6226a00c53bd0ea70ed4d33d2fa9affbfec3cb10f9c48b7b12eb5fe384d6" Mar 13 09:45:28 crc kubenswrapper[4930]: I0313 09:45:28.864841 4930 scope.go:117] "RemoveContainer" containerID="9a3ed9cb0a0840f8467dee30175cffc85b41d9dd0e5e058647f8d4d1745926bd" Mar 13 09:45:28 crc kubenswrapper[4930]: I0313 09:45:28.898524 4930 scope.go:117] "RemoveContainer" containerID="c4033b1c387e1729a85fe822e90af40e36d600af1237c3dd5bfa0f7a2e713838" Mar 13 09:45:28 crc kubenswrapper[4930]: I0313 09:45:28.928484 4930 scope.go:117] "RemoveContainer" containerID="7e38480a43a6e1adcd53ab5a8cfc002475af7a65aca6c38215a2af9876233bc2" Mar 13 09:45:28 crc kubenswrapper[4930]: I0313 09:45:28.965565 4930 scope.go:117] "RemoveContainer" containerID="681d0cd0f264d27cd661797cc1039f195824ca91fbb3eac1e6d8422a9da38dc1" Mar 13 09:45:28 crc kubenswrapper[4930]: I0313 09:45:28.995938 4930 scope.go:117] "RemoveContainer" containerID="dd293d0f1b807793ae505f3f997c54e9ecb690d62a8ec6da2b542b321eb582ab" Mar 13 09:45:29 crc kubenswrapper[4930]: I0313 09:45:29.025180 4930 scope.go:117] "RemoveContainer" containerID="afa291e400ac690b8ecca6419a12980d873fe93d54b442559582f30cd0bf61fb" Mar 13 09:45:32 crc kubenswrapper[4930]: I0313 09:45:32.075927 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-mqk65"] Mar 13 09:45:32 crc kubenswrapper[4930]: I0313 09:45:32.097058 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-mqk65"] Mar 13 09:45:32 crc kubenswrapper[4930]: I0313 09:45:32.109973 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-djnqj"] Mar 13 09:45:32 crc kubenswrapper[4930]: I0313 09:45:32.121225 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-djnqj"] Mar 13 09:45:33 crc kubenswrapper[4930]: I0313 09:45:33.985796 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10e08b2d-bcbf-4b39-9baa-e4ca5213567c" path="/var/lib/kubelet/pods/10e08b2d-bcbf-4b39-9baa-e4ca5213567c/volumes" Mar 13 09:45:33 crc kubenswrapper[4930]: I0313 09:45:33.987781 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66e9286f-5c5e-49e5-9952-8d01782f2a09" path="/var/lib/kubelet/pods/66e9286f-5c5e-49e5-9952-8d01782f2a09/volumes" Mar 13 09:45:36 crc kubenswrapper[4930]: I0313 09:45:36.047267 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-h9bf6"] Mar 13 09:45:36 crc kubenswrapper[4930]: I0313 09:45:36.058136 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-h9bf6"] Mar 13 09:45:37 crc kubenswrapper[4930]: I0313 09:45:37.991719 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3fb676a-c7df-4e03-a4ad-41da78bde051" path="/var/lib/kubelet/pods/d3fb676a-c7df-4e03-a4ad-41da78bde051/volumes" Mar 13 09:45:46 crc kubenswrapper[4930]: I0313 09:45:46.959350 4930 generic.go:334] "Generic (PLEG): container finished" podID="3331ecec-adf5-4107-a891-07d2fd6e4be7" containerID="fc8b94642ed37d8e6adc9b63f377e6cee6aea018e3f794e0ea1d2f64fc82ea58" exitCode=0 Mar 13 09:45:46 crc kubenswrapper[4930]: I0313 09:45:46.959484 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" event={"ID":"3331ecec-adf5-4107-a891-07d2fd6e4be7","Type":"ContainerDied","Data":"fc8b94642ed37d8e6adc9b63f377e6cee6aea018e3f794e0ea1d2f64fc82ea58"} Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.441691 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.602820 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3331ecec-adf5-4107-a891-07d2fd6e4be7-ssh-key-openstack-edpm-ipam\") pod \"3331ecec-adf5-4107-a891-07d2fd6e4be7\" (UID: \"3331ecec-adf5-4107-a891-07d2fd6e4be7\") " Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.602896 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpf27\" (UniqueName: \"kubernetes.io/projected/3331ecec-adf5-4107-a891-07d2fd6e4be7-kube-api-access-cpf27\") pod \"3331ecec-adf5-4107-a891-07d2fd6e4be7\" (UID: \"3331ecec-adf5-4107-a891-07d2fd6e4be7\") " Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.603064 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3331ecec-adf5-4107-a891-07d2fd6e4be7-inventory\") pod \"3331ecec-adf5-4107-a891-07d2fd6e4be7\" (UID: \"3331ecec-adf5-4107-a891-07d2fd6e4be7\") " Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.609514 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3331ecec-adf5-4107-a891-07d2fd6e4be7-kube-api-access-cpf27" (OuterVolumeSpecName: "kube-api-access-cpf27") pod "3331ecec-adf5-4107-a891-07d2fd6e4be7" (UID: "3331ecec-adf5-4107-a891-07d2fd6e4be7"). InnerVolumeSpecName "kube-api-access-cpf27". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.638808 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3331ecec-adf5-4107-a891-07d2fd6e4be7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3331ecec-adf5-4107-a891-07d2fd6e4be7" (UID: "3331ecec-adf5-4107-a891-07d2fd6e4be7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.641087 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3331ecec-adf5-4107-a891-07d2fd6e4be7-inventory" (OuterVolumeSpecName: "inventory") pod "3331ecec-adf5-4107-a891-07d2fd6e4be7" (UID: "3331ecec-adf5-4107-a891-07d2fd6e4be7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.706117 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3331ecec-adf5-4107-a891-07d2fd6e4be7-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.706154 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3331ecec-adf5-4107-a891-07d2fd6e4be7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.706167 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpf27\" (UniqueName: \"kubernetes.io/projected/3331ecec-adf5-4107-a891-07d2fd6e4be7-kube-api-access-cpf27\") on node \"crc\" DevicePath \"\"" Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.979598 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" event={"ID":"3331ecec-adf5-4107-a891-07d2fd6e4be7","Type":"ContainerDied","Data":"5c759cf6072d6eae031caeffbf59585ab2a9319e53341657dfd8bdf739870246"} Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.980012 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c759cf6072d6eae031caeffbf59585ab2a9319e53341657dfd8bdf739870246" Mar 13 09:45:48 crc kubenswrapper[4930]: I0313 09:45:48.979691 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-wvc2m" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.072778 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz"] Mar 13 09:45:49 crc kubenswrapper[4930]: E0313 09:45:49.073649 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed8cfe62-57c8-42c3-82bb-d712ff3f8265" containerName="collect-profiles" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.073737 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed8cfe62-57c8-42c3-82bb-d712ff3f8265" containerName="collect-profiles" Mar 13 09:45:49 crc kubenswrapper[4930]: E0313 09:45:49.073881 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3331ecec-adf5-4107-a891-07d2fd6e4be7" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.073947 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="3331ecec-adf5-4107-a891-07d2fd6e4be7" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.074293 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed8cfe62-57c8-42c3-82bb-d712ff3f8265" containerName="collect-profiles" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.074388 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="3331ecec-adf5-4107-a891-07d2fd6e4be7" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.075503 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.078395 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.079259 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.080344 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.080527 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.110456 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz"] Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.117981 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b9930ef-bf59-4b1c-b66a-e46198319285-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz\" (UID: \"3b9930ef-bf59-4b1c-b66a-e46198319285\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.118166 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd6k2\" (UniqueName: \"kubernetes.io/projected/3b9930ef-bf59-4b1c-b66a-e46198319285-kube-api-access-vd6k2\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz\" (UID: \"3b9930ef-bf59-4b1c-b66a-e46198319285\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.118219 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3b9930ef-bf59-4b1c-b66a-e46198319285-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz\" (UID: \"3b9930ef-bf59-4b1c-b66a-e46198319285\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.220075 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b9930ef-bf59-4b1c-b66a-e46198319285-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz\" (UID: \"3b9930ef-bf59-4b1c-b66a-e46198319285\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.220212 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd6k2\" (UniqueName: \"kubernetes.io/projected/3b9930ef-bf59-4b1c-b66a-e46198319285-kube-api-access-vd6k2\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz\" (UID: \"3b9930ef-bf59-4b1c-b66a-e46198319285\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.220239 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3b9930ef-bf59-4b1c-b66a-e46198319285-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz\" (UID: \"3b9930ef-bf59-4b1c-b66a-e46198319285\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.224412 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3b9930ef-bf59-4b1c-b66a-e46198319285-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz\" (UID: \"3b9930ef-bf59-4b1c-b66a-e46198319285\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.224885 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b9930ef-bf59-4b1c-b66a-e46198319285-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz\" (UID: \"3b9930ef-bf59-4b1c-b66a-e46198319285\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.240187 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd6k2\" (UniqueName: \"kubernetes.io/projected/3b9930ef-bf59-4b1c-b66a-e46198319285-kube-api-access-vd6k2\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz\" (UID: \"3b9930ef-bf59-4b1c-b66a-e46198319285\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:45:49 crc kubenswrapper[4930]: I0313 09:45:49.424284 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:45:50 crc kubenswrapper[4930]: I0313 09:45:50.083718 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz"] Mar 13 09:45:50 crc kubenswrapper[4930]: I0313 09:45:50.089526 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 09:45:51 crc kubenswrapper[4930]: I0313 09:45:51.007331 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" event={"ID":"3b9930ef-bf59-4b1c-b66a-e46198319285","Type":"ContainerStarted","Data":"a9a47da6f248115085f524e81e4cb1620aa51632701f0f48368bd3a50fb698ba"} Mar 13 09:45:51 crc kubenswrapper[4930]: I0313 09:45:51.007964 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" event={"ID":"3b9930ef-bf59-4b1c-b66a-e46198319285","Type":"ContainerStarted","Data":"8abbbd8a17836d411f23eef8cea327577897532750c98d83a70270e85cbcf4a7"} Mar 13 09:45:51 crc kubenswrapper[4930]: I0313 09:45:51.038680 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" podStartSLOduration=1.563667905 podStartE2EDuration="2.038665082s" podCreationTimestamp="2026-03-13 09:45:49 +0000 UTC" firstStartedPulling="2026-03-13 09:45:50.089281573 +0000 UTC m=+1990.839196240" lastFinishedPulling="2026-03-13 09:45:50.56427874 +0000 UTC m=+1991.314193417" observedRunningTime="2026-03-13 09:45:51.036929918 +0000 UTC m=+1991.786844635" watchObservedRunningTime="2026-03-13 09:45:51.038665082 +0000 UTC m=+1991.788579759" Mar 13 09:45:52 crc kubenswrapper[4930]: I0313 09:45:52.040016 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-7dsrx"] Mar 13 09:45:52 crc kubenswrapper[4930]: I0313 09:45:52.054792 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-7dsrx"] Mar 13 09:45:53 crc kubenswrapper[4930]: I0313 09:45:53.986361 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98c022d1-9a39-4f3e-8108-e4be2b287077" path="/var/lib/kubelet/pods/98c022d1-9a39-4f3e-8108-e4be2b287077/volumes" Mar 13 09:46:00 crc kubenswrapper[4930]: I0313 09:46:00.138575 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556586-v4txf"] Mar 13 09:46:00 crc kubenswrapper[4930]: I0313 09:46:00.141284 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556586-v4txf" Mar 13 09:46:00 crc kubenswrapper[4930]: I0313 09:46:00.143500 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:46:00 crc kubenswrapper[4930]: I0313 09:46:00.143826 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:46:00 crc kubenswrapper[4930]: I0313 09:46:00.143870 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:46:00 crc kubenswrapper[4930]: I0313 09:46:00.161923 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556586-v4txf"] Mar 13 09:46:00 crc kubenswrapper[4930]: I0313 09:46:00.282704 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vbrq6\" (UniqueName: \"kubernetes.io/projected/3cc4b291-7e47-46d6-b8f2-400d377bb246-kube-api-access-vbrq6\") pod \"auto-csr-approver-29556586-v4txf\" (UID: \"3cc4b291-7e47-46d6-b8f2-400d377bb246\") " pod="openshift-infra/auto-csr-approver-29556586-v4txf" Mar 13 09:46:00 crc kubenswrapper[4930]: I0313 09:46:00.384887 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vbrq6\" (UniqueName: \"kubernetes.io/projected/3cc4b291-7e47-46d6-b8f2-400d377bb246-kube-api-access-vbrq6\") pod \"auto-csr-approver-29556586-v4txf\" (UID: \"3cc4b291-7e47-46d6-b8f2-400d377bb246\") " pod="openshift-infra/auto-csr-approver-29556586-v4txf" Mar 13 09:46:00 crc kubenswrapper[4930]: I0313 09:46:00.402889 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vbrq6\" (UniqueName: \"kubernetes.io/projected/3cc4b291-7e47-46d6-b8f2-400d377bb246-kube-api-access-vbrq6\") pod \"auto-csr-approver-29556586-v4txf\" (UID: \"3cc4b291-7e47-46d6-b8f2-400d377bb246\") " pod="openshift-infra/auto-csr-approver-29556586-v4txf" Mar 13 09:46:00 crc kubenswrapper[4930]: I0313 09:46:00.460893 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556586-v4txf" Mar 13 09:46:00 crc kubenswrapper[4930]: I0313 09:46:00.950948 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556586-v4txf"] Mar 13 09:46:01 crc kubenswrapper[4930]: I0313 09:46:01.123238 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556586-v4txf" event={"ID":"3cc4b291-7e47-46d6-b8f2-400d377bb246","Type":"ContainerStarted","Data":"4eddc10345b710a1e71772bb70bdac64bb9fadeb69635270f592303bee374f35"} Mar 13 09:46:03 crc kubenswrapper[4930]: I0313 09:46:03.147074 4930 generic.go:334] "Generic (PLEG): container finished" podID="3cc4b291-7e47-46d6-b8f2-400d377bb246" containerID="5a718ce2e52c8d0da7055d8d9a7108271368e58fe64d9f5c97880b28f9030acf" exitCode=0 Mar 13 09:46:03 crc kubenswrapper[4930]: I0313 09:46:03.147152 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556586-v4txf" event={"ID":"3cc4b291-7e47-46d6-b8f2-400d377bb246","Type":"ContainerDied","Data":"5a718ce2e52c8d0da7055d8d9a7108271368e58fe64d9f5c97880b28f9030acf"} Mar 13 09:46:04 crc kubenswrapper[4930]: I0313 09:46:04.566883 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556586-v4txf" Mar 13 09:46:04 crc kubenswrapper[4930]: I0313 09:46:04.692598 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vbrq6\" (UniqueName: \"kubernetes.io/projected/3cc4b291-7e47-46d6-b8f2-400d377bb246-kube-api-access-vbrq6\") pod \"3cc4b291-7e47-46d6-b8f2-400d377bb246\" (UID: \"3cc4b291-7e47-46d6-b8f2-400d377bb246\") " Mar 13 09:46:04 crc kubenswrapper[4930]: I0313 09:46:04.697918 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cc4b291-7e47-46d6-b8f2-400d377bb246-kube-api-access-vbrq6" (OuterVolumeSpecName: "kube-api-access-vbrq6") pod "3cc4b291-7e47-46d6-b8f2-400d377bb246" (UID: "3cc4b291-7e47-46d6-b8f2-400d377bb246"). InnerVolumeSpecName "kube-api-access-vbrq6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:46:04 crc kubenswrapper[4930]: I0313 09:46:04.796482 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vbrq6\" (UniqueName: \"kubernetes.io/projected/3cc4b291-7e47-46d6-b8f2-400d377bb246-kube-api-access-vbrq6\") on node \"crc\" DevicePath \"\"" Mar 13 09:46:05 crc kubenswrapper[4930]: I0313 09:46:05.170405 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556586-v4txf" event={"ID":"3cc4b291-7e47-46d6-b8f2-400d377bb246","Type":"ContainerDied","Data":"4eddc10345b710a1e71772bb70bdac64bb9fadeb69635270f592303bee374f35"} Mar 13 09:46:05 crc kubenswrapper[4930]: I0313 09:46:05.170488 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4eddc10345b710a1e71772bb70bdac64bb9fadeb69635270f592303bee374f35" Mar 13 09:46:05 crc kubenswrapper[4930]: I0313 09:46:05.170512 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556586-v4txf" Mar 13 09:46:05 crc kubenswrapper[4930]: I0313 09:46:05.647922 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556580-xvwz8"] Mar 13 09:46:05 crc kubenswrapper[4930]: I0313 09:46:05.661884 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556580-xvwz8"] Mar 13 09:46:06 crc kubenswrapper[4930]: I0313 09:46:06.018899 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d55c277-21b0-4e8f-a561-c20ba8c2ce8d" path="/var/lib/kubelet/pods/3d55c277-21b0-4e8f-a561-c20ba8c2ce8d/volumes" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.110835 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-x28d8"] Mar 13 09:46:10 crc kubenswrapper[4930]: E0313 09:46:10.111643 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cc4b291-7e47-46d6-b8f2-400d377bb246" containerName="oc" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.111658 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cc4b291-7e47-46d6-b8f2-400d377bb246" containerName="oc" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.111918 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cc4b291-7e47-46d6-b8f2-400d377bb246" containerName="oc" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.113537 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.126119 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x28d8"] Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.229748 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/379bb1a4-6575-45bd-9c97-81b861cd6e89-catalog-content\") pod \"community-operators-x28d8\" (UID: \"379bb1a4-6575-45bd-9c97-81b861cd6e89\") " pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.229818 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9c42x\" (UniqueName: \"kubernetes.io/projected/379bb1a4-6575-45bd-9c97-81b861cd6e89-kube-api-access-9c42x\") pod \"community-operators-x28d8\" (UID: \"379bb1a4-6575-45bd-9c97-81b861cd6e89\") " pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.230197 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/379bb1a4-6575-45bd-9c97-81b861cd6e89-utilities\") pod \"community-operators-x28d8\" (UID: \"379bb1a4-6575-45bd-9c97-81b861cd6e89\") " pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.332344 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/379bb1a4-6575-45bd-9c97-81b861cd6e89-utilities\") pod \"community-operators-x28d8\" (UID: \"379bb1a4-6575-45bd-9c97-81b861cd6e89\") " pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.332623 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/379bb1a4-6575-45bd-9c97-81b861cd6e89-catalog-content\") pod \"community-operators-x28d8\" (UID: \"379bb1a4-6575-45bd-9c97-81b861cd6e89\") " pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.332689 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9c42x\" (UniqueName: \"kubernetes.io/projected/379bb1a4-6575-45bd-9c97-81b861cd6e89-kube-api-access-9c42x\") pod \"community-operators-x28d8\" (UID: \"379bb1a4-6575-45bd-9c97-81b861cd6e89\") " pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.332894 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/379bb1a4-6575-45bd-9c97-81b861cd6e89-utilities\") pod \"community-operators-x28d8\" (UID: \"379bb1a4-6575-45bd-9c97-81b861cd6e89\") " pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.333185 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/379bb1a4-6575-45bd-9c97-81b861cd6e89-catalog-content\") pod \"community-operators-x28d8\" (UID: \"379bb1a4-6575-45bd-9c97-81b861cd6e89\") " pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.364341 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9c42x\" (UniqueName: \"kubernetes.io/projected/379bb1a4-6575-45bd-9c97-81b861cd6e89-kube-api-access-9c42x\") pod \"community-operators-x28d8\" (UID: \"379bb1a4-6575-45bd-9c97-81b861cd6e89\") " pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:10 crc kubenswrapper[4930]: I0313 09:46:10.437182 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:11 crc kubenswrapper[4930]: I0313 09:46:11.078114 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-x28d8"] Mar 13 09:46:11 crc kubenswrapper[4930]: W0313 09:46:11.085557 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod379bb1a4_6575_45bd_9c97_81b861cd6e89.slice/crio-b4646ce3c29d8b9bc506ab1c87da1b8df91a575a224d7346d5fb26d0ef532350 WatchSource:0}: Error finding container b4646ce3c29d8b9bc506ab1c87da1b8df91a575a224d7346d5fb26d0ef532350: Status 404 returned error can't find the container with id b4646ce3c29d8b9bc506ab1c87da1b8df91a575a224d7346d5fb26d0ef532350 Mar 13 09:46:11 crc kubenswrapper[4930]: I0313 09:46:11.242973 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x28d8" event={"ID":"379bb1a4-6575-45bd-9c97-81b861cd6e89","Type":"ContainerStarted","Data":"b4646ce3c29d8b9bc506ab1c87da1b8df91a575a224d7346d5fb26d0ef532350"} Mar 13 09:46:12 crc kubenswrapper[4930]: I0313 09:46:12.256469 4930 generic.go:334] "Generic (PLEG): container finished" podID="379bb1a4-6575-45bd-9c97-81b861cd6e89" containerID="54efe9133229a70bc53a8d9ecb99ae61f5b2530d98167b8802a0ffce3b5d22e4" exitCode=0 Mar 13 09:46:12 crc kubenswrapper[4930]: I0313 09:46:12.256566 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x28d8" event={"ID":"379bb1a4-6575-45bd-9c97-81b861cd6e89","Type":"ContainerDied","Data":"54efe9133229a70bc53a8d9ecb99ae61f5b2530d98167b8802a0ffce3b5d22e4"} Mar 13 09:46:12 crc kubenswrapper[4930]: I0313 09:46:12.308317 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:46:12 crc kubenswrapper[4930]: I0313 09:46:12.308387 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:46:13 crc kubenswrapper[4930]: I0313 09:46:13.271407 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x28d8" event={"ID":"379bb1a4-6575-45bd-9c97-81b861cd6e89","Type":"ContainerStarted","Data":"26a6b4dd4efb20d5b5173e9af9f30c8ea405f5f73094f1911991325c8c624e53"} Mar 13 09:46:15 crc kubenswrapper[4930]: I0313 09:46:15.293089 4930 generic.go:334] "Generic (PLEG): container finished" podID="379bb1a4-6575-45bd-9c97-81b861cd6e89" containerID="26a6b4dd4efb20d5b5173e9af9f30c8ea405f5f73094f1911991325c8c624e53" exitCode=0 Mar 13 09:46:15 crc kubenswrapper[4930]: I0313 09:46:15.293178 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x28d8" event={"ID":"379bb1a4-6575-45bd-9c97-81b861cd6e89","Type":"ContainerDied","Data":"26a6b4dd4efb20d5b5173e9af9f30c8ea405f5f73094f1911991325c8c624e53"} Mar 13 09:46:16 crc kubenswrapper[4930]: I0313 09:46:16.306911 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x28d8" event={"ID":"379bb1a4-6575-45bd-9c97-81b861cd6e89","Type":"ContainerStarted","Data":"831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2"} Mar 13 09:46:16 crc kubenswrapper[4930]: I0313 09:46:16.331731 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-x28d8" podStartSLOduration=2.918213792 podStartE2EDuration="6.331703573s" podCreationTimestamp="2026-03-13 09:46:10 +0000 UTC" firstStartedPulling="2026-03-13 09:46:12.25891849 +0000 UTC m=+2013.008833167" lastFinishedPulling="2026-03-13 09:46:15.672408271 +0000 UTC m=+2016.422322948" observedRunningTime="2026-03-13 09:46:16.330653997 +0000 UTC m=+2017.080568724" watchObservedRunningTime="2026-03-13 09:46:16.331703573 +0000 UTC m=+2017.081618260" Mar 13 09:46:20 crc kubenswrapper[4930]: I0313 09:46:20.437534 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:20 crc kubenswrapper[4930]: I0313 09:46:20.438070 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:21 crc kubenswrapper[4930]: I0313 09:46:21.485681 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-x28d8" podUID="379bb1a4-6575-45bd-9c97-81b861cd6e89" containerName="registry-server" probeResult="failure" output=< Mar 13 09:46:21 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:46:21 crc kubenswrapper[4930]: > Mar 13 09:46:29 crc kubenswrapper[4930]: I0313 09:46:29.275477 4930 scope.go:117] "RemoveContainer" containerID="3c1f9e12a3d90e441e6432e6727e9aa72a5dac133603645f94a85129d8be0a90" Mar 13 09:46:29 crc kubenswrapper[4930]: I0313 09:46:29.323786 4930 scope.go:117] "RemoveContainer" containerID="d6e741c1a755fc3772e85e18f098c7f5930be5a84a595c19c2f9249b2409cc75" Mar 13 09:46:29 crc kubenswrapper[4930]: I0313 09:46:29.392138 4930 scope.go:117] "RemoveContainer" containerID="c959109492edd2baef172e975037522c981f72caba726ee20c38339cc13fd18d" Mar 13 09:46:29 crc kubenswrapper[4930]: I0313 09:46:29.452327 4930 scope.go:117] "RemoveContainer" containerID="ac95d58ba906ac3a4c40acd96be960d3b3dfc8dacf73f640c3c779d6f7b101df" Mar 13 09:46:29 crc kubenswrapper[4930]: I0313 09:46:29.513482 4930 scope.go:117] "RemoveContainer" containerID="cb71701e4d8a02633aae086604550da618517bad71a9e78fdb788997e04e82c5" Mar 13 09:46:30 crc kubenswrapper[4930]: I0313 09:46:30.483593 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:30 crc kubenswrapper[4930]: I0313 09:46:30.545373 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:30 crc kubenswrapper[4930]: I0313 09:46:30.726172 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x28d8"] Mar 13 09:46:32 crc kubenswrapper[4930]: I0313 09:46:32.471490 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-x28d8" podUID="379bb1a4-6575-45bd-9c97-81b861cd6e89" containerName="registry-server" containerID="cri-o://831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2" gracePeriod=2 Mar 13 09:46:32 crc kubenswrapper[4930]: I0313 09:46:32.975696 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.103935 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9c42x\" (UniqueName: \"kubernetes.io/projected/379bb1a4-6575-45bd-9c97-81b861cd6e89-kube-api-access-9c42x\") pod \"379bb1a4-6575-45bd-9c97-81b861cd6e89\" (UID: \"379bb1a4-6575-45bd-9c97-81b861cd6e89\") " Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.104005 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/379bb1a4-6575-45bd-9c97-81b861cd6e89-catalog-content\") pod \"379bb1a4-6575-45bd-9c97-81b861cd6e89\" (UID: \"379bb1a4-6575-45bd-9c97-81b861cd6e89\") " Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.104178 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/379bb1a4-6575-45bd-9c97-81b861cd6e89-utilities\") pod \"379bb1a4-6575-45bd-9c97-81b861cd6e89\" (UID: \"379bb1a4-6575-45bd-9c97-81b861cd6e89\") " Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.105331 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/379bb1a4-6575-45bd-9c97-81b861cd6e89-utilities" (OuterVolumeSpecName: "utilities") pod "379bb1a4-6575-45bd-9c97-81b861cd6e89" (UID: "379bb1a4-6575-45bd-9c97-81b861cd6e89"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.110055 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/379bb1a4-6575-45bd-9c97-81b861cd6e89-kube-api-access-9c42x" (OuterVolumeSpecName: "kube-api-access-9c42x") pod "379bb1a4-6575-45bd-9c97-81b861cd6e89" (UID: "379bb1a4-6575-45bd-9c97-81b861cd6e89"). InnerVolumeSpecName "kube-api-access-9c42x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.170630 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/379bb1a4-6575-45bd-9c97-81b861cd6e89-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "379bb1a4-6575-45bd-9c97-81b861cd6e89" (UID: "379bb1a4-6575-45bd-9c97-81b861cd6e89"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.206932 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/379bb1a4-6575-45bd-9c97-81b861cd6e89-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.206973 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9c42x\" (UniqueName: \"kubernetes.io/projected/379bb1a4-6575-45bd-9c97-81b861cd6e89-kube-api-access-9c42x\") on node \"crc\" DevicePath \"\"" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.206987 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/379bb1a4-6575-45bd-9c97-81b861cd6e89-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.482948 4930 generic.go:334] "Generic (PLEG): container finished" podID="379bb1a4-6575-45bd-9c97-81b861cd6e89" containerID="831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2" exitCode=0 Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.483003 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x28d8" event={"ID":"379bb1a4-6575-45bd-9c97-81b861cd6e89","Type":"ContainerDied","Data":"831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2"} Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.483041 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-x28d8" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.483084 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-x28d8" event={"ID":"379bb1a4-6575-45bd-9c97-81b861cd6e89","Type":"ContainerDied","Data":"b4646ce3c29d8b9bc506ab1c87da1b8df91a575a224d7346d5fb26d0ef532350"} Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.483112 4930 scope.go:117] "RemoveContainer" containerID="831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.514745 4930 scope.go:117] "RemoveContainer" containerID="26a6b4dd4efb20d5b5173e9af9f30c8ea405f5f73094f1911991325c8c624e53" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.529195 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-x28d8"] Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.539613 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-x28d8"] Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.545735 4930 scope.go:117] "RemoveContainer" containerID="54efe9133229a70bc53a8d9ecb99ae61f5b2530d98167b8802a0ffce3b5d22e4" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.617455 4930 scope.go:117] "RemoveContainer" containerID="831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2" Mar 13 09:46:33 crc kubenswrapper[4930]: E0313 09:46:33.617945 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2\": container with ID starting with 831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2 not found: ID does not exist" containerID="831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.618079 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2"} err="failed to get container status \"831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2\": rpc error: code = NotFound desc = could not find container \"831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2\": container with ID starting with 831c65e43f2b856b4ac1b04178f57b4c0a3c6a265f59fd65242f0d3f84b169b2 not found: ID does not exist" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.618200 4930 scope.go:117] "RemoveContainer" containerID="26a6b4dd4efb20d5b5173e9af9f30c8ea405f5f73094f1911991325c8c624e53" Mar 13 09:46:33 crc kubenswrapper[4930]: E0313 09:46:33.618643 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26a6b4dd4efb20d5b5173e9af9f30c8ea405f5f73094f1911991325c8c624e53\": container with ID starting with 26a6b4dd4efb20d5b5173e9af9f30c8ea405f5f73094f1911991325c8c624e53 not found: ID does not exist" containerID="26a6b4dd4efb20d5b5173e9af9f30c8ea405f5f73094f1911991325c8c624e53" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.618665 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26a6b4dd4efb20d5b5173e9af9f30c8ea405f5f73094f1911991325c8c624e53"} err="failed to get container status \"26a6b4dd4efb20d5b5173e9af9f30c8ea405f5f73094f1911991325c8c624e53\": rpc error: code = NotFound desc = could not find container \"26a6b4dd4efb20d5b5173e9af9f30c8ea405f5f73094f1911991325c8c624e53\": container with ID starting with 26a6b4dd4efb20d5b5173e9af9f30c8ea405f5f73094f1911991325c8c624e53 not found: ID does not exist" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.618679 4930 scope.go:117] "RemoveContainer" containerID="54efe9133229a70bc53a8d9ecb99ae61f5b2530d98167b8802a0ffce3b5d22e4" Mar 13 09:46:33 crc kubenswrapper[4930]: E0313 09:46:33.618920 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54efe9133229a70bc53a8d9ecb99ae61f5b2530d98167b8802a0ffce3b5d22e4\": container with ID starting with 54efe9133229a70bc53a8d9ecb99ae61f5b2530d98167b8802a0ffce3b5d22e4 not found: ID does not exist" containerID="54efe9133229a70bc53a8d9ecb99ae61f5b2530d98167b8802a0ffce3b5d22e4" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.618942 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54efe9133229a70bc53a8d9ecb99ae61f5b2530d98167b8802a0ffce3b5d22e4"} err="failed to get container status \"54efe9133229a70bc53a8d9ecb99ae61f5b2530d98167b8802a0ffce3b5d22e4\": rpc error: code = NotFound desc = could not find container \"54efe9133229a70bc53a8d9ecb99ae61f5b2530d98167b8802a0ffce3b5d22e4\": container with ID starting with 54efe9133229a70bc53a8d9ecb99ae61f5b2530d98167b8802a0ffce3b5d22e4 not found: ID does not exist" Mar 13 09:46:33 crc kubenswrapper[4930]: I0313 09:46:33.982548 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="379bb1a4-6575-45bd-9c97-81b861cd6e89" path="/var/lib/kubelet/pods/379bb1a4-6575-45bd-9c97-81b861cd6e89/volumes" Mar 13 09:46:42 crc kubenswrapper[4930]: I0313 09:46:42.308228 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:46:42 crc kubenswrapper[4930]: I0313 09:46:42.308809 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:46:48 crc kubenswrapper[4930]: I0313 09:46:48.070043 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-122f-account-create-update-9988g"] Mar 13 09:46:48 crc kubenswrapper[4930]: I0313 09:46:48.081362 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-122f-account-create-update-9988g"] Mar 13 09:46:49 crc kubenswrapper[4930]: I0313 09:46:49.038035 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-qbssl"] Mar 13 09:46:49 crc kubenswrapper[4930]: I0313 09:46:49.051962 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-xxmsm"] Mar 13 09:46:49 crc kubenswrapper[4930]: I0313 09:46:49.071274 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-qbssl"] Mar 13 09:46:49 crc kubenswrapper[4930]: I0313 09:46:49.087841 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-xxmsm"] Mar 13 09:46:50 crc kubenswrapper[4930]: I0313 09:46:50.000408 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3c2e237-7f4c-48a2-8a8d-f5838192decc" path="/var/lib/kubelet/pods/b3c2e237-7f4c-48a2-8a8d-f5838192decc/volumes" Mar 13 09:46:50 crc kubenswrapper[4930]: I0313 09:46:50.002590 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ec86dec2-eb5a-497e-bba2-d1b96d80c65c" path="/var/lib/kubelet/pods/ec86dec2-eb5a-497e-bba2-d1b96d80c65c/volumes" Mar 13 09:46:50 crc kubenswrapper[4930]: I0313 09:46:50.003578 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5942b4a-ef6d-40ca-b67f-e7965513004d" path="/var/lib/kubelet/pods/f5942b4a-ef6d-40ca-b67f-e7965513004d/volumes" Mar 13 09:46:50 crc kubenswrapper[4930]: I0313 09:46:50.049615 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-7e6b-account-create-update-tgmkk"] Mar 13 09:46:50 crc kubenswrapper[4930]: I0313 09:46:50.065211 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-7e6b-account-create-update-tgmkk"] Mar 13 09:46:51 crc kubenswrapper[4930]: I0313 09:46:51.071688 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-b5f3-account-create-update-9zsnn"] Mar 13 09:46:51 crc kubenswrapper[4930]: I0313 09:46:51.085963 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-lzrnq"] Mar 13 09:46:51 crc kubenswrapper[4930]: I0313 09:46:51.096521 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-b5f3-account-create-update-9zsnn"] Mar 13 09:46:51 crc kubenswrapper[4930]: I0313 09:46:51.107505 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-lzrnq"] Mar 13 09:46:51 crc kubenswrapper[4930]: I0313 09:46:51.664396 4930 generic.go:334] "Generic (PLEG): container finished" podID="3b9930ef-bf59-4b1c-b66a-e46198319285" containerID="a9a47da6f248115085f524e81e4cb1620aa51632701f0f48368bd3a50fb698ba" exitCode=0 Mar 13 09:46:51 crc kubenswrapper[4930]: I0313 09:46:51.664482 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" event={"ID":"3b9930ef-bf59-4b1c-b66a-e46198319285","Type":"ContainerDied","Data":"a9a47da6f248115085f524e81e4cb1620aa51632701f0f48368bd3a50fb698ba"} Mar 13 09:46:51 crc kubenswrapper[4930]: I0313 09:46:51.993304 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a15fab1-c914-4555-813a-ab9552b29a22" path="/var/lib/kubelet/pods/4a15fab1-c914-4555-813a-ab9552b29a22/volumes" Mar 13 09:46:52 crc kubenswrapper[4930]: I0313 09:46:52.007660 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4a92113b-21a5-4db3-9708-589fc0aa57e8" path="/var/lib/kubelet/pods/4a92113b-21a5-4db3-9708-589fc0aa57e8/volumes" Mar 13 09:46:52 crc kubenswrapper[4930]: I0313 09:46:52.009912 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e12cb85-3947-4c60-96a6-8ac07f5e63e8" path="/var/lib/kubelet/pods/9e12cb85-3947-4c60-96a6-8ac07f5e63e8/volumes" Mar 13 09:46:52 crc kubenswrapper[4930]: I0313 09:46:52.942375 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-znxl9"] Mar 13 09:46:52 crc kubenswrapper[4930]: E0313 09:46:52.943501 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="379bb1a4-6575-45bd-9c97-81b861cd6e89" containerName="registry-server" Mar 13 09:46:52 crc kubenswrapper[4930]: I0313 09:46:52.943516 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="379bb1a4-6575-45bd-9c97-81b861cd6e89" containerName="registry-server" Mar 13 09:46:52 crc kubenswrapper[4930]: E0313 09:46:52.943558 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="379bb1a4-6575-45bd-9c97-81b861cd6e89" containerName="extract-utilities" Mar 13 09:46:52 crc kubenswrapper[4930]: I0313 09:46:52.943567 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="379bb1a4-6575-45bd-9c97-81b861cd6e89" containerName="extract-utilities" Mar 13 09:46:52 crc kubenswrapper[4930]: E0313 09:46:52.943597 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="379bb1a4-6575-45bd-9c97-81b861cd6e89" containerName="extract-content" Mar 13 09:46:52 crc kubenswrapper[4930]: I0313 09:46:52.943628 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="379bb1a4-6575-45bd-9c97-81b861cd6e89" containerName="extract-content" Mar 13 09:46:52 crc kubenswrapper[4930]: I0313 09:46:52.944245 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="379bb1a4-6575-45bd-9c97-81b861cd6e89" containerName="registry-server" Mar 13 09:46:52 crc kubenswrapper[4930]: I0313 09:46:52.947971 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:46:52 crc kubenswrapper[4930]: I0313 09:46:52.959453 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-znxl9"] Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.037363 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6vqn\" (UniqueName: \"kubernetes.io/projected/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-kube-api-access-t6vqn\") pod \"redhat-operators-znxl9\" (UID: \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\") " pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.037842 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-catalog-content\") pod \"redhat-operators-znxl9\" (UID: \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\") " pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.038011 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-utilities\") pod \"redhat-operators-znxl9\" (UID: \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\") " pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.140994 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6vqn\" (UniqueName: \"kubernetes.io/projected/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-kube-api-access-t6vqn\") pod \"redhat-operators-znxl9\" (UID: \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\") " pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.141093 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-catalog-content\") pod \"redhat-operators-znxl9\" (UID: \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\") " pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.141161 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-utilities\") pod \"redhat-operators-znxl9\" (UID: \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\") " pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.142095 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-utilities\") pod \"redhat-operators-znxl9\" (UID: \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\") " pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.142630 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-catalog-content\") pod \"redhat-operators-znxl9\" (UID: \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\") " pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.160257 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6vqn\" (UniqueName: \"kubernetes.io/projected/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-kube-api-access-t6vqn\") pod \"redhat-operators-znxl9\" (UID: \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\") " pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.254071 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.278310 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.345414 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b9930ef-bf59-4b1c-b66a-e46198319285-inventory\") pod \"3b9930ef-bf59-4b1c-b66a-e46198319285\" (UID: \"3b9930ef-bf59-4b1c-b66a-e46198319285\") " Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.345606 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd6k2\" (UniqueName: \"kubernetes.io/projected/3b9930ef-bf59-4b1c-b66a-e46198319285-kube-api-access-vd6k2\") pod \"3b9930ef-bf59-4b1c-b66a-e46198319285\" (UID: \"3b9930ef-bf59-4b1c-b66a-e46198319285\") " Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.345756 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3b9930ef-bf59-4b1c-b66a-e46198319285-ssh-key-openstack-edpm-ipam\") pod \"3b9930ef-bf59-4b1c-b66a-e46198319285\" (UID: \"3b9930ef-bf59-4b1c-b66a-e46198319285\") " Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.351143 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b9930ef-bf59-4b1c-b66a-e46198319285-kube-api-access-vd6k2" (OuterVolumeSpecName: "kube-api-access-vd6k2") pod "3b9930ef-bf59-4b1c-b66a-e46198319285" (UID: "3b9930ef-bf59-4b1c-b66a-e46198319285"). InnerVolumeSpecName "kube-api-access-vd6k2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.380760 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b9930ef-bf59-4b1c-b66a-e46198319285-inventory" (OuterVolumeSpecName: "inventory") pod "3b9930ef-bf59-4b1c-b66a-e46198319285" (UID: "3b9930ef-bf59-4b1c-b66a-e46198319285"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.381777 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b9930ef-bf59-4b1c-b66a-e46198319285-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3b9930ef-bf59-4b1c-b66a-e46198319285" (UID: "3b9930ef-bf59-4b1c-b66a-e46198319285"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.449156 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3b9930ef-bf59-4b1c-b66a-e46198319285-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.449189 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3b9930ef-bf59-4b1c-b66a-e46198319285-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.449198 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd6k2\" (UniqueName: \"kubernetes.io/projected/3b9930ef-bf59-4b1c-b66a-e46198319285-kube-api-access-vd6k2\") on node \"crc\" DevicePath \"\"" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.689705 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" event={"ID":"3b9930ef-bf59-4b1c-b66a-e46198319285","Type":"ContainerDied","Data":"8abbbd8a17836d411f23eef8cea327577897532750c98d83a70270e85cbcf4a7"} Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.690084 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8abbbd8a17836d411f23eef8cea327577897532750c98d83a70270e85cbcf4a7" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.689826 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vf5wz" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.770806 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-znxl9"] Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.808568 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82"] Mar 13 09:46:53 crc kubenswrapper[4930]: E0313 09:46:53.809246 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b9930ef-bf59-4b1c-b66a-e46198319285" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.809271 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b9930ef-bf59-4b1c-b66a-e46198319285" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.809506 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b9930ef-bf59-4b1c-b66a-e46198319285" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.810323 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.813524 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.813529 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.813682 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.819554 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.836691 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82"] Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.858942 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ffwmz\" (UniqueName: \"kubernetes.io/projected/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-kube-api-access-ffwmz\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bkk82\" (UID: \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.859037 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bkk82\" (UID: \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.859186 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bkk82\" (UID: \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.963190 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ffwmz\" (UniqueName: \"kubernetes.io/projected/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-kube-api-access-ffwmz\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bkk82\" (UID: \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.963269 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bkk82\" (UID: \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.963349 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bkk82\" (UID: \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.971157 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bkk82\" (UID: \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.973108 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bkk82\" (UID: \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:46:53 crc kubenswrapper[4930]: I0313 09:46:53.989989 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ffwmz\" (UniqueName: \"kubernetes.io/projected/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-kube-api-access-ffwmz\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-bkk82\" (UID: \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:46:54 crc kubenswrapper[4930]: I0313 09:46:54.139325 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:46:54 crc kubenswrapper[4930]: I0313 09:46:54.705341 4930 generic.go:334] "Generic (PLEG): container finished" podID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerID="6d635ecdbf2605476aa7074a8649b88571ca452e49aa29416c583924c1094f88" exitCode=0 Mar 13 09:46:54 crc kubenswrapper[4930]: I0313 09:46:54.705541 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znxl9" event={"ID":"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba","Type":"ContainerDied","Data":"6d635ecdbf2605476aa7074a8649b88571ca452e49aa29416c583924c1094f88"} Mar 13 09:46:54 crc kubenswrapper[4930]: I0313 09:46:54.705725 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znxl9" event={"ID":"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba","Type":"ContainerStarted","Data":"171953a3a482849ba0e8b6c933875627236ea9eb93b068a9910a1f7f1f096639"} Mar 13 09:46:54 crc kubenswrapper[4930]: I0313 09:46:54.783068 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82"] Mar 13 09:46:55 crc kubenswrapper[4930]: I0313 09:46:55.719798 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" event={"ID":"01cb69ce-fe5d-49d3-bae0-a5a332ecf910","Type":"ContainerStarted","Data":"757252f0d92d92bdb6fcc69286801abf1520d1669b0fc9ddc10445a55fbf73cc"} Mar 13 09:46:55 crc kubenswrapper[4930]: I0313 09:46:55.720598 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" event={"ID":"01cb69ce-fe5d-49d3-bae0-a5a332ecf910","Type":"ContainerStarted","Data":"1bf31ac8527ee5a3c6da6b59ca31ea232ef2b042c85d85fb7f5780e0527a68db"} Mar 13 09:46:55 crc kubenswrapper[4930]: I0313 09:46:55.754023 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" podStartSLOduration=2.307775687 podStartE2EDuration="2.753987761s" podCreationTimestamp="2026-03-13 09:46:53 +0000 UTC" firstStartedPulling="2026-03-13 09:46:54.788945521 +0000 UTC m=+2055.538860198" lastFinishedPulling="2026-03-13 09:46:55.235157595 +0000 UTC m=+2055.985072272" observedRunningTime="2026-03-13 09:46:55.742589708 +0000 UTC m=+2056.492504385" watchObservedRunningTime="2026-03-13 09:46:55.753987761 +0000 UTC m=+2056.503902448" Mar 13 09:46:56 crc kubenswrapper[4930]: I0313 09:46:56.736641 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znxl9" event={"ID":"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba","Type":"ContainerStarted","Data":"02c07201f8a360a7d637e292a2393c346f502f939ad4d69f9f238db4acf6f917"} Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.307785 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6jjdf"] Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.310450 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.319414 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jjdf"] Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.410081 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfxxj\" (UniqueName: \"kubernetes.io/projected/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-kube-api-access-kfxxj\") pod \"redhat-marketplace-6jjdf\" (UID: \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\") " pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.410455 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-utilities\") pod \"redhat-marketplace-6jjdf\" (UID: \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\") " pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.410599 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-catalog-content\") pod \"redhat-marketplace-6jjdf\" (UID: \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\") " pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.513673 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfxxj\" (UniqueName: \"kubernetes.io/projected/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-kube-api-access-kfxxj\") pod \"redhat-marketplace-6jjdf\" (UID: \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\") " pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.513768 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-utilities\") pod \"redhat-marketplace-6jjdf\" (UID: \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\") " pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.513793 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-catalog-content\") pod \"redhat-marketplace-6jjdf\" (UID: \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\") " pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.514603 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-catalog-content\") pod \"redhat-marketplace-6jjdf\" (UID: \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\") " pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.514665 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-utilities\") pod \"redhat-marketplace-6jjdf\" (UID: \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\") " pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.544007 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfxxj\" (UniqueName: \"kubernetes.io/projected/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-kube-api-access-kfxxj\") pod \"redhat-marketplace-6jjdf\" (UID: \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\") " pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:46:58 crc kubenswrapper[4930]: I0313 09:46:58.639448 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:46:59 crc kubenswrapper[4930]: I0313 09:46:59.321247 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jjdf"] Mar 13 09:46:59 crc kubenswrapper[4930]: I0313 09:46:59.777989 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jjdf" event={"ID":"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006","Type":"ContainerStarted","Data":"4d16d587ec467eabcc5bf22119259a3b7bfc3a2bf53e50b5bffafc02b3dfb8c0"} Mar 13 09:47:00 crc kubenswrapper[4930]: I0313 09:47:00.791086 4930 generic.go:334] "Generic (PLEG): container finished" podID="01cb69ce-fe5d-49d3-bae0-a5a332ecf910" containerID="757252f0d92d92bdb6fcc69286801abf1520d1669b0fc9ddc10445a55fbf73cc" exitCode=0 Mar 13 09:47:00 crc kubenswrapper[4930]: I0313 09:47:00.791166 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" event={"ID":"01cb69ce-fe5d-49d3-bae0-a5a332ecf910","Type":"ContainerDied","Data":"757252f0d92d92bdb6fcc69286801abf1520d1669b0fc9ddc10445a55fbf73cc"} Mar 13 09:47:00 crc kubenswrapper[4930]: I0313 09:47:00.792569 4930 generic.go:334] "Generic (PLEG): container finished" podID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" containerID="db5a7e5405cb8268ef48bc49ae420ae2f71babfcb118095942f4cb3d49340548" exitCode=0 Mar 13 09:47:00 crc kubenswrapper[4930]: I0313 09:47:00.792604 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jjdf" event={"ID":"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006","Type":"ContainerDied","Data":"db5a7e5405cb8268ef48bc49ae420ae2f71babfcb118095942f4cb3d49340548"} Mar 13 09:47:01 crc kubenswrapper[4930]: I0313 09:47:01.824838 4930 generic.go:334] "Generic (PLEG): container finished" podID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerID="02c07201f8a360a7d637e292a2393c346f502f939ad4d69f9f238db4acf6f917" exitCode=0 Mar 13 09:47:01 crc kubenswrapper[4930]: I0313 09:47:01.824889 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znxl9" event={"ID":"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba","Type":"ContainerDied","Data":"02c07201f8a360a7d637e292a2393c346f502f939ad4d69f9f238db4acf6f917"} Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.495708 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.552559 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-inventory\") pod \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\" (UID: \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\") " Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.552649 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ffwmz\" (UniqueName: \"kubernetes.io/projected/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-kube-api-access-ffwmz\") pod \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\" (UID: \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\") " Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.552740 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-ssh-key-openstack-edpm-ipam\") pod \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\" (UID: \"01cb69ce-fe5d-49d3-bae0-a5a332ecf910\") " Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.569965 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-kube-api-access-ffwmz" (OuterVolumeSpecName: "kube-api-access-ffwmz") pod "01cb69ce-fe5d-49d3-bae0-a5a332ecf910" (UID: "01cb69ce-fe5d-49d3-bae0-a5a332ecf910"). InnerVolumeSpecName "kube-api-access-ffwmz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.622360 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-inventory" (OuterVolumeSpecName: "inventory") pod "01cb69ce-fe5d-49d3-bae0-a5a332ecf910" (UID: "01cb69ce-fe5d-49d3-bae0-a5a332ecf910"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.632282 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "01cb69ce-fe5d-49d3-bae0-a5a332ecf910" (UID: "01cb69ce-fe5d-49d3-bae0-a5a332ecf910"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.655139 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.655175 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ffwmz\" (UniqueName: \"kubernetes.io/projected/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-kube-api-access-ffwmz\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.655189 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/01cb69ce-fe5d-49d3-bae0-a5a332ecf910-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.852568 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" event={"ID":"01cb69ce-fe5d-49d3-bae0-a5a332ecf910","Type":"ContainerDied","Data":"1bf31ac8527ee5a3c6da6b59ca31ea232ef2b042c85d85fb7f5780e0527a68db"} Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.852618 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1bf31ac8527ee5a3c6da6b59ca31ea232ef2b042c85d85fb7f5780e0527a68db" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.852586 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-bkk82" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.856201 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znxl9" event={"ID":"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba","Type":"ContainerStarted","Data":"053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c"} Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.862424 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jjdf" event={"ID":"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006","Type":"ContainerStarted","Data":"aa27697a3b17f16ceded788cf77e12787c2b58fe531eb9a459100b225a5e8d9c"} Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.916797 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-znxl9" podStartSLOduration=3.27646056 podStartE2EDuration="10.916772717s" podCreationTimestamp="2026-03-13 09:46:52 +0000 UTC" firstStartedPulling="2026-03-13 09:46:54.708500935 +0000 UTC m=+2055.458415612" lastFinishedPulling="2026-03-13 09:47:02.348813092 +0000 UTC m=+2063.098727769" observedRunningTime="2026-03-13 09:47:02.872932489 +0000 UTC m=+2063.622847166" watchObservedRunningTime="2026-03-13 09:47:02.916772717 +0000 UTC m=+2063.666687404" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.993700 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz"] Mar 13 09:47:02 crc kubenswrapper[4930]: E0313 09:47:02.994329 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01cb69ce-fe5d-49d3-bae0-a5a332ecf910" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.994556 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="01cb69ce-fe5d-49d3-bae0-a5a332ecf910" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.995027 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="01cb69ce-fe5d-49d3-bae0-a5a332ecf910" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.996128 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.998035 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.998035 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:47:02 crc kubenswrapper[4930]: I0313 09:47:02.998479 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.000932 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.006406 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz"] Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.167927 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/44e80226-1db6-4df7-9561-eb33be2f35ca-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wflfz\" (UID: \"44e80226-1db6-4df7-9561-eb33be2f35ca\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.168708 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44e80226-1db6-4df7-9561-eb33be2f35ca-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wflfz\" (UID: \"44e80226-1db6-4df7-9561-eb33be2f35ca\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.168809 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7ttdf\" (UniqueName: \"kubernetes.io/projected/44e80226-1db6-4df7-9561-eb33be2f35ca-kube-api-access-7ttdf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wflfz\" (UID: \"44e80226-1db6-4df7-9561-eb33be2f35ca\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.271201 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44e80226-1db6-4df7-9561-eb33be2f35ca-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wflfz\" (UID: \"44e80226-1db6-4df7-9561-eb33be2f35ca\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.271291 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7ttdf\" (UniqueName: \"kubernetes.io/projected/44e80226-1db6-4df7-9561-eb33be2f35ca-kube-api-access-7ttdf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wflfz\" (UID: \"44e80226-1db6-4df7-9561-eb33be2f35ca\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.271375 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/44e80226-1db6-4df7-9561-eb33be2f35ca-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wflfz\" (UID: \"44e80226-1db6-4df7-9561-eb33be2f35ca\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.276988 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/44e80226-1db6-4df7-9561-eb33be2f35ca-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wflfz\" (UID: \"44e80226-1db6-4df7-9561-eb33be2f35ca\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.277064 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44e80226-1db6-4df7-9561-eb33be2f35ca-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wflfz\" (UID: \"44e80226-1db6-4df7-9561-eb33be2f35ca\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.278748 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.278783 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.294669 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7ttdf\" (UniqueName: \"kubernetes.io/projected/44e80226-1db6-4df7-9561-eb33be2f35ca-kube-api-access-7ttdf\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-wflfz\" (UID: \"44e80226-1db6-4df7-9561-eb33be2f35ca\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:03 crc kubenswrapper[4930]: I0313 09:47:03.317594 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:04 crc kubenswrapper[4930]: I0313 09:47:04.012728 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz"] Mar 13 09:47:04 crc kubenswrapper[4930]: I0313 09:47:04.335588 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-znxl9" podUID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerName="registry-server" probeResult="failure" output=< Mar 13 09:47:04 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:47:04 crc kubenswrapper[4930]: > Mar 13 09:47:04 crc kubenswrapper[4930]: I0313 09:47:04.919825 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" event={"ID":"44e80226-1db6-4df7-9561-eb33be2f35ca","Type":"ContainerStarted","Data":"2c84a94862a7497ea8131f967eb44790202a792ba94101cb6b55e33f7e5eaa52"} Mar 13 09:47:04 crc kubenswrapper[4930]: I0313 09:47:04.920352 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" event={"ID":"44e80226-1db6-4df7-9561-eb33be2f35ca","Type":"ContainerStarted","Data":"9d77fd63df0c020a4a7d5a12a58917d8e312d3e3e853c47aef01638d0b53a551"} Mar 13 09:47:04 crc kubenswrapper[4930]: I0313 09:47:04.922713 4930 generic.go:334] "Generic (PLEG): container finished" podID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" containerID="aa27697a3b17f16ceded788cf77e12787c2b58fe531eb9a459100b225a5e8d9c" exitCode=0 Mar 13 09:47:04 crc kubenswrapper[4930]: I0313 09:47:04.922771 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jjdf" event={"ID":"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006","Type":"ContainerDied","Data":"aa27697a3b17f16ceded788cf77e12787c2b58fe531eb9a459100b225a5e8d9c"} Mar 13 09:47:04 crc kubenswrapper[4930]: I0313 09:47:04.938973 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" podStartSLOduration=2.533968531 podStartE2EDuration="2.938925189s" podCreationTimestamp="2026-03-13 09:47:02 +0000 UTC" firstStartedPulling="2026-03-13 09:47:04.014583071 +0000 UTC m=+2064.764497748" lastFinishedPulling="2026-03-13 09:47:04.419539729 +0000 UTC m=+2065.169454406" observedRunningTime="2026-03-13 09:47:04.937127264 +0000 UTC m=+2065.687041941" watchObservedRunningTime="2026-03-13 09:47:04.938925189 +0000 UTC m=+2065.688839866" Mar 13 09:47:05 crc kubenswrapper[4930]: I0313 09:47:05.937834 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jjdf" event={"ID":"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006","Type":"ContainerStarted","Data":"eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01"} Mar 13 09:47:05 crc kubenswrapper[4930]: I0313 09:47:05.957892 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6jjdf" podStartSLOduration=3.403175555 podStartE2EDuration="7.957876636s" podCreationTimestamp="2026-03-13 09:46:58 +0000 UTC" firstStartedPulling="2026-03-13 09:47:00.794870879 +0000 UTC m=+2061.544785556" lastFinishedPulling="2026-03-13 09:47:05.34957196 +0000 UTC m=+2066.099486637" observedRunningTime="2026-03-13 09:47:05.956211635 +0000 UTC m=+2066.706126312" watchObservedRunningTime="2026-03-13 09:47:05.957876636 +0000 UTC m=+2066.707791313" Mar 13 09:47:08 crc kubenswrapper[4930]: I0313 09:47:08.640346 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:47:08 crc kubenswrapper[4930]: I0313 09:47:08.641181 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:47:09 crc kubenswrapper[4930]: I0313 09:47:09.693107 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-6jjdf" podUID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" containerName="registry-server" probeResult="failure" output=< Mar 13 09:47:09 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:47:09 crc kubenswrapper[4930]: > Mar 13 09:47:12 crc kubenswrapper[4930]: I0313 09:47:12.308884 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:47:12 crc kubenswrapper[4930]: I0313 09:47:12.309136 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:47:12 crc kubenswrapper[4930]: I0313 09:47:12.309182 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:47:12 crc kubenswrapper[4930]: I0313 09:47:12.310140 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9d9e501a7dfec9be2a4a7aae030dcc1a1a33117fc6ee6f36debdcb82b8cfee89"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 09:47:12 crc kubenswrapper[4930]: I0313 09:47:12.310475 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://9d9e501a7dfec9be2a4a7aae030dcc1a1a33117fc6ee6f36debdcb82b8cfee89" gracePeriod=600 Mar 13 09:47:13 crc kubenswrapper[4930]: I0313 09:47:13.028700 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="9d9e501a7dfec9be2a4a7aae030dcc1a1a33117fc6ee6f36debdcb82b8cfee89" exitCode=0 Mar 13 09:47:13 crc kubenswrapper[4930]: I0313 09:47:13.028740 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"9d9e501a7dfec9be2a4a7aae030dcc1a1a33117fc6ee6f36debdcb82b8cfee89"} Mar 13 09:47:13 crc kubenswrapper[4930]: I0313 09:47:13.029048 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f"} Mar 13 09:47:13 crc kubenswrapper[4930]: I0313 09:47:13.029081 4930 scope.go:117] "RemoveContainer" containerID="638e39a288a00b8d558f1dfe278cd5e13514c51b47801b2f707aaeab30037766" Mar 13 09:47:14 crc kubenswrapper[4930]: I0313 09:47:14.331566 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-znxl9" podUID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerName="registry-server" probeResult="failure" output=< Mar 13 09:47:14 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 09:47:14 crc kubenswrapper[4930]: > Mar 13 09:47:18 crc kubenswrapper[4930]: I0313 09:47:18.692066 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:47:18 crc kubenswrapper[4930]: I0313 09:47:18.743076 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:47:18 crc kubenswrapper[4930]: I0313 09:47:18.943351 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jjdf"] Mar 13 09:47:20 crc kubenswrapper[4930]: I0313 09:47:20.117842 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6jjdf" podUID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" containerName="registry-server" containerID="cri-o://eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01" gracePeriod=2 Mar 13 09:47:20 crc kubenswrapper[4930]: I0313 09:47:20.722157 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:47:20 crc kubenswrapper[4930]: I0313 09:47:20.805417 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-catalog-content\") pod \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\" (UID: \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\") " Mar 13 09:47:20 crc kubenswrapper[4930]: I0313 09:47:20.805666 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-utilities\") pod \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\" (UID: \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\") " Mar 13 09:47:20 crc kubenswrapper[4930]: I0313 09:47:20.805775 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfxxj\" (UniqueName: \"kubernetes.io/projected/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-kube-api-access-kfxxj\") pod \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\" (UID: \"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006\") " Mar 13 09:47:20 crc kubenswrapper[4930]: I0313 09:47:20.806228 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-utilities" (OuterVolumeSpecName: "utilities") pod "94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" (UID: "94bcc27a-a620-41fd-b8e0-e4e7d4a2f006"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:47:20 crc kubenswrapper[4930]: I0313 09:47:20.806930 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:20 crc kubenswrapper[4930]: I0313 09:47:20.827665 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-kube-api-access-kfxxj" (OuterVolumeSpecName: "kube-api-access-kfxxj") pod "94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" (UID: "94bcc27a-a620-41fd-b8e0-e4e7d4a2f006"). InnerVolumeSpecName "kube-api-access-kfxxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:47:20 crc kubenswrapper[4930]: I0313 09:47:20.834580 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" (UID: "94bcc27a-a620-41fd-b8e0-e4e7d4a2f006"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:47:20 crc kubenswrapper[4930]: I0313 09:47:20.908955 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfxxj\" (UniqueName: \"kubernetes.io/projected/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-kube-api-access-kfxxj\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:20 crc kubenswrapper[4930]: I0313 09:47:20.908988 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.129926 4930 generic.go:334] "Generic (PLEG): container finished" podID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" containerID="eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01" exitCode=0 Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.129980 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6jjdf" Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.129998 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jjdf" event={"ID":"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006","Type":"ContainerDied","Data":"eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01"} Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.131278 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6jjdf" event={"ID":"94bcc27a-a620-41fd-b8e0-e4e7d4a2f006","Type":"ContainerDied","Data":"4d16d587ec467eabcc5bf22119259a3b7bfc3a2bf53e50b5bffafc02b3dfb8c0"} Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.131302 4930 scope.go:117] "RemoveContainer" containerID="eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01" Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.151993 4930 scope.go:117] "RemoveContainer" containerID="aa27697a3b17f16ceded788cf77e12787c2b58fe531eb9a459100b225a5e8d9c" Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.171660 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jjdf"] Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.180206 4930 scope.go:117] "RemoveContainer" containerID="db5a7e5405cb8268ef48bc49ae420ae2f71babfcb118095942f4cb3d49340548" Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.182411 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6jjdf"] Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.239231 4930 scope.go:117] "RemoveContainer" containerID="eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01" Mar 13 09:47:21 crc kubenswrapper[4930]: E0313 09:47:21.239640 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01\": container with ID starting with eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01 not found: ID does not exist" containerID="eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01" Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.239684 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01"} err="failed to get container status \"eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01\": rpc error: code = NotFound desc = could not find container \"eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01\": container with ID starting with eeb4045ffd86f2715b3dd6feb222c5000060016a7ea5f147456a020b1a212d01 not found: ID does not exist" Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.239716 4930 scope.go:117] "RemoveContainer" containerID="aa27697a3b17f16ceded788cf77e12787c2b58fe531eb9a459100b225a5e8d9c" Mar 13 09:47:21 crc kubenswrapper[4930]: E0313 09:47:21.240001 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa27697a3b17f16ceded788cf77e12787c2b58fe531eb9a459100b225a5e8d9c\": container with ID starting with aa27697a3b17f16ceded788cf77e12787c2b58fe531eb9a459100b225a5e8d9c not found: ID does not exist" containerID="aa27697a3b17f16ceded788cf77e12787c2b58fe531eb9a459100b225a5e8d9c" Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.240025 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa27697a3b17f16ceded788cf77e12787c2b58fe531eb9a459100b225a5e8d9c"} err="failed to get container status \"aa27697a3b17f16ceded788cf77e12787c2b58fe531eb9a459100b225a5e8d9c\": rpc error: code = NotFound desc = could not find container \"aa27697a3b17f16ceded788cf77e12787c2b58fe531eb9a459100b225a5e8d9c\": container with ID starting with aa27697a3b17f16ceded788cf77e12787c2b58fe531eb9a459100b225a5e8d9c not found: ID does not exist" Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.240041 4930 scope.go:117] "RemoveContainer" containerID="db5a7e5405cb8268ef48bc49ae420ae2f71babfcb118095942f4cb3d49340548" Mar 13 09:47:21 crc kubenswrapper[4930]: E0313 09:47:21.240269 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db5a7e5405cb8268ef48bc49ae420ae2f71babfcb118095942f4cb3d49340548\": container with ID starting with db5a7e5405cb8268ef48bc49ae420ae2f71babfcb118095942f4cb3d49340548 not found: ID does not exist" containerID="db5a7e5405cb8268ef48bc49ae420ae2f71babfcb118095942f4cb3d49340548" Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.240315 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db5a7e5405cb8268ef48bc49ae420ae2f71babfcb118095942f4cb3d49340548"} err="failed to get container status \"db5a7e5405cb8268ef48bc49ae420ae2f71babfcb118095942f4cb3d49340548\": rpc error: code = NotFound desc = could not find container \"db5a7e5405cb8268ef48bc49ae420ae2f71babfcb118095942f4cb3d49340548\": container with ID starting with db5a7e5405cb8268ef48bc49ae420ae2f71babfcb118095942f4cb3d49340548 not found: ID does not exist" Mar 13 09:47:21 crc kubenswrapper[4930]: I0313 09:47:21.993288 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" path="/var/lib/kubelet/pods/94bcc27a-a620-41fd-b8e0-e4e7d4a2f006/volumes" Mar 13 09:47:22 crc kubenswrapper[4930]: I0313 09:47:22.047802 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gb9wl"] Mar 13 09:47:22 crc kubenswrapper[4930]: I0313 09:47:22.065588 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-gb9wl"] Mar 13 09:47:23 crc kubenswrapper[4930]: I0313 09:47:23.342312 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:47:23 crc kubenswrapper[4930]: I0313 09:47:23.393110 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:47:23 crc kubenswrapper[4930]: I0313 09:47:23.985168 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f2a8264a-7145-4f0b-9f7e-4e9552dc84c0" path="/var/lib/kubelet/pods/f2a8264a-7145-4f0b-9f7e-4e9552dc84c0/volumes" Mar 13 09:47:24 crc kubenswrapper[4930]: I0313 09:47:24.334053 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-znxl9"] Mar 13 09:47:25 crc kubenswrapper[4930]: I0313 09:47:25.170531 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-znxl9" podUID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerName="registry-server" containerID="cri-o://053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c" gracePeriod=2 Mar 13 09:47:25 crc kubenswrapper[4930]: I0313 09:47:25.770496 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:47:25 crc kubenswrapper[4930]: I0313 09:47:25.867165 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-utilities\") pod \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\" (UID: \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\") " Mar 13 09:47:25 crc kubenswrapper[4930]: I0313 09:47:25.867354 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-catalog-content\") pod \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\" (UID: \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\") " Mar 13 09:47:25 crc kubenswrapper[4930]: I0313 09:47:25.867632 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6vqn\" (UniqueName: \"kubernetes.io/projected/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-kube-api-access-t6vqn\") pod \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\" (UID: \"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba\") " Mar 13 09:47:25 crc kubenswrapper[4930]: I0313 09:47:25.868035 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-utilities" (OuterVolumeSpecName: "utilities") pod "482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" (UID: "482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:47:25 crc kubenswrapper[4930]: I0313 09:47:25.868996 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:25 crc kubenswrapper[4930]: I0313 09:47:25.874770 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-kube-api-access-t6vqn" (OuterVolumeSpecName: "kube-api-access-t6vqn") pod "482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" (UID: "482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba"). InnerVolumeSpecName "kube-api-access-t6vqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:47:25 crc kubenswrapper[4930]: I0313 09:47:25.971171 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6vqn\" (UniqueName: \"kubernetes.io/projected/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-kube-api-access-t6vqn\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.009790 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" (UID: "482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.073235 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.198478 4930 generic.go:334] "Generic (PLEG): container finished" podID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerID="053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c" exitCode=0 Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.198533 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znxl9" event={"ID":"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba","Type":"ContainerDied","Data":"053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c"} Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.198568 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-znxl9" event={"ID":"482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba","Type":"ContainerDied","Data":"171953a3a482849ba0e8b6c933875627236ea9eb93b068a9910a1f7f1f096639"} Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.198567 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-znxl9" Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.198589 4930 scope.go:117] "RemoveContainer" containerID="053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c" Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.233334 4930 scope.go:117] "RemoveContainer" containerID="02c07201f8a360a7d637e292a2393c346f502f939ad4d69f9f238db4acf6f917" Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.251542 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-znxl9"] Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.263032 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-znxl9"] Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.268798 4930 scope.go:117] "RemoveContainer" containerID="6d635ecdbf2605476aa7074a8649b88571ca452e49aa29416c583924c1094f88" Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.326984 4930 scope.go:117] "RemoveContainer" containerID="053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c" Mar 13 09:47:26 crc kubenswrapper[4930]: E0313 09:47:26.327464 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c\": container with ID starting with 053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c not found: ID does not exist" containerID="053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c" Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.327501 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c"} err="failed to get container status \"053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c\": rpc error: code = NotFound desc = could not find container \"053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c\": container with ID starting with 053a3663c34c12c92e72c73de516de72ea8c8093822b809fa779b9bfde6d476c not found: ID does not exist" Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.327524 4930 scope.go:117] "RemoveContainer" containerID="02c07201f8a360a7d637e292a2393c346f502f939ad4d69f9f238db4acf6f917" Mar 13 09:47:26 crc kubenswrapper[4930]: E0313 09:47:26.328003 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02c07201f8a360a7d637e292a2393c346f502f939ad4d69f9f238db4acf6f917\": container with ID starting with 02c07201f8a360a7d637e292a2393c346f502f939ad4d69f9f238db4acf6f917 not found: ID does not exist" containerID="02c07201f8a360a7d637e292a2393c346f502f939ad4d69f9f238db4acf6f917" Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.328040 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02c07201f8a360a7d637e292a2393c346f502f939ad4d69f9f238db4acf6f917"} err="failed to get container status \"02c07201f8a360a7d637e292a2393c346f502f939ad4d69f9f238db4acf6f917\": rpc error: code = NotFound desc = could not find container \"02c07201f8a360a7d637e292a2393c346f502f939ad4d69f9f238db4acf6f917\": container with ID starting with 02c07201f8a360a7d637e292a2393c346f502f939ad4d69f9f238db4acf6f917 not found: ID does not exist" Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.328053 4930 scope.go:117] "RemoveContainer" containerID="6d635ecdbf2605476aa7074a8649b88571ca452e49aa29416c583924c1094f88" Mar 13 09:47:26 crc kubenswrapper[4930]: E0313 09:47:26.328384 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d635ecdbf2605476aa7074a8649b88571ca452e49aa29416c583924c1094f88\": container with ID starting with 6d635ecdbf2605476aa7074a8649b88571ca452e49aa29416c583924c1094f88 not found: ID does not exist" containerID="6d635ecdbf2605476aa7074a8649b88571ca452e49aa29416c583924c1094f88" Mar 13 09:47:26 crc kubenswrapper[4930]: I0313 09:47:26.328413 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d635ecdbf2605476aa7074a8649b88571ca452e49aa29416c583924c1094f88"} err="failed to get container status \"6d635ecdbf2605476aa7074a8649b88571ca452e49aa29416c583924c1094f88\": rpc error: code = NotFound desc = could not find container \"6d635ecdbf2605476aa7074a8649b88571ca452e49aa29416c583924c1094f88\": container with ID starting with 6d635ecdbf2605476aa7074a8649b88571ca452e49aa29416c583924c1094f88 not found: ID does not exist" Mar 13 09:47:27 crc kubenswrapper[4930]: I0313 09:47:27.984383 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" path="/var/lib/kubelet/pods/482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba/volumes" Mar 13 09:47:29 crc kubenswrapper[4930]: I0313 09:47:29.670494 4930 scope.go:117] "RemoveContainer" containerID="ddbb33d16390e104959af69b3bda2520dd6ea5437de7ac7f96f224286260373f" Mar 13 09:47:29 crc kubenswrapper[4930]: I0313 09:47:29.703415 4930 scope.go:117] "RemoveContainer" containerID="4b7b5aa15cd29fc57e05fac737b0fb42c5eb2c5caaab5c9fa05a061b5338ee86" Mar 13 09:47:29 crc kubenswrapper[4930]: I0313 09:47:29.756714 4930 scope.go:117] "RemoveContainer" containerID="2fd9d220e0ef2e9c5e49e551039bcf3e92dafde12abc470bdf438161092241d1" Mar 13 09:47:29 crc kubenswrapper[4930]: I0313 09:47:29.820648 4930 scope.go:117] "RemoveContainer" containerID="8eb6dd8aa8cf56e87cf75580b57bbd99dd58ff1bac13350056a5a725df6555b4" Mar 13 09:47:29 crc kubenswrapper[4930]: I0313 09:47:29.862793 4930 scope.go:117] "RemoveContainer" containerID="fc112d9419344d327e1c8c8a6349dd19da83a7b1e827b7fef330b29325fea0da" Mar 13 09:47:29 crc kubenswrapper[4930]: I0313 09:47:29.923669 4930 scope.go:117] "RemoveContainer" containerID="eb1cc95c17cab0807e9c358ba5604ad7f4577cbc6877f1993d79c9e79ea03e29" Mar 13 09:47:29 crc kubenswrapper[4930]: I0313 09:47:29.966682 4930 scope.go:117] "RemoveContainer" containerID="93396c58766bfe9d2b827cfca22cf86170091e1e938e713952dfa6b7ed7f9479" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.048353 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7nzdh"] Mar 13 09:47:39 crc kubenswrapper[4930]: E0313 09:47:39.049575 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" containerName="registry-server" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.049598 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" containerName="registry-server" Mar 13 09:47:39 crc kubenswrapper[4930]: E0313 09:47:39.049623 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerName="extract-content" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.049630 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerName="extract-content" Mar 13 09:47:39 crc kubenswrapper[4930]: E0313 09:47:39.049647 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" containerName="extract-utilities" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.049654 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" containerName="extract-utilities" Mar 13 09:47:39 crc kubenswrapper[4930]: E0313 09:47:39.049667 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerName="registry-server" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.049673 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerName="registry-server" Mar 13 09:47:39 crc kubenswrapper[4930]: E0313 09:47:39.049696 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" containerName="extract-content" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.049704 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" containerName="extract-content" Mar 13 09:47:39 crc kubenswrapper[4930]: E0313 09:47:39.049718 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerName="extract-utilities" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.049727 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerName="extract-utilities" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.050405 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="482b6b4d-0353-4cc2-ba5b-18cbc7ec51ba" containerName="registry-server" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.050458 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="94bcc27a-a620-41fd-b8e0-e4e7d4a2f006" containerName="registry-server" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.052239 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.063233 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7nzdh"] Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.206308 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2914b248-f67d-49a8-b908-8fed6028f477-utilities\") pod \"certified-operators-7nzdh\" (UID: \"2914b248-f67d-49a8-b908-8fed6028f477\") " pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.206509 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pd6j\" (UniqueName: \"kubernetes.io/projected/2914b248-f67d-49a8-b908-8fed6028f477-kube-api-access-8pd6j\") pod \"certified-operators-7nzdh\" (UID: \"2914b248-f67d-49a8-b908-8fed6028f477\") " pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.206591 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2914b248-f67d-49a8-b908-8fed6028f477-catalog-content\") pod \"certified-operators-7nzdh\" (UID: \"2914b248-f67d-49a8-b908-8fed6028f477\") " pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.308957 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2914b248-f67d-49a8-b908-8fed6028f477-utilities\") pod \"certified-operators-7nzdh\" (UID: \"2914b248-f67d-49a8-b908-8fed6028f477\") " pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.309082 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pd6j\" (UniqueName: \"kubernetes.io/projected/2914b248-f67d-49a8-b908-8fed6028f477-kube-api-access-8pd6j\") pod \"certified-operators-7nzdh\" (UID: \"2914b248-f67d-49a8-b908-8fed6028f477\") " pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.309177 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2914b248-f67d-49a8-b908-8fed6028f477-catalog-content\") pod \"certified-operators-7nzdh\" (UID: \"2914b248-f67d-49a8-b908-8fed6028f477\") " pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.309551 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2914b248-f67d-49a8-b908-8fed6028f477-utilities\") pod \"certified-operators-7nzdh\" (UID: \"2914b248-f67d-49a8-b908-8fed6028f477\") " pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.309918 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2914b248-f67d-49a8-b908-8fed6028f477-catalog-content\") pod \"certified-operators-7nzdh\" (UID: \"2914b248-f67d-49a8-b908-8fed6028f477\") " pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.328283 4930 generic.go:334] "Generic (PLEG): container finished" podID="44e80226-1db6-4df7-9561-eb33be2f35ca" containerID="2c84a94862a7497ea8131f967eb44790202a792ba94101cb6b55e33f7e5eaa52" exitCode=0 Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.328358 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" event={"ID":"44e80226-1db6-4df7-9561-eb33be2f35ca","Type":"ContainerDied","Data":"2c84a94862a7497ea8131f967eb44790202a792ba94101cb6b55e33f7e5eaa52"} Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.331530 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pd6j\" (UniqueName: \"kubernetes.io/projected/2914b248-f67d-49a8-b908-8fed6028f477-kube-api-access-8pd6j\") pod \"certified-operators-7nzdh\" (UID: \"2914b248-f67d-49a8-b908-8fed6028f477\") " pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.373628 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:39 crc kubenswrapper[4930]: I0313 09:47:39.839773 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7nzdh"] Mar 13 09:47:40 crc kubenswrapper[4930]: I0313 09:47:40.340018 4930 generic.go:334] "Generic (PLEG): container finished" podID="2914b248-f67d-49a8-b908-8fed6028f477" containerID="4f1dd9a3b8991c95b91d81a584b59d04a6e71ee611d88a5cb38e1451872b14f2" exitCode=0 Mar 13 09:47:40 crc kubenswrapper[4930]: I0313 09:47:40.340115 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nzdh" event={"ID":"2914b248-f67d-49a8-b908-8fed6028f477","Type":"ContainerDied","Data":"4f1dd9a3b8991c95b91d81a584b59d04a6e71ee611d88a5cb38e1451872b14f2"} Mar 13 09:47:40 crc kubenswrapper[4930]: I0313 09:47:40.340387 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nzdh" event={"ID":"2914b248-f67d-49a8-b908-8fed6028f477","Type":"ContainerStarted","Data":"39f64ceac38f2d1c4f4b92038f243c6417c46be970ff20e094e310d10ab55a2e"} Mar 13 09:47:40 crc kubenswrapper[4930]: I0313 09:47:40.947748 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.068881 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7ttdf\" (UniqueName: \"kubernetes.io/projected/44e80226-1db6-4df7-9561-eb33be2f35ca-kube-api-access-7ttdf\") pod \"44e80226-1db6-4df7-9561-eb33be2f35ca\" (UID: \"44e80226-1db6-4df7-9561-eb33be2f35ca\") " Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.069108 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/44e80226-1db6-4df7-9561-eb33be2f35ca-ssh-key-openstack-edpm-ipam\") pod \"44e80226-1db6-4df7-9561-eb33be2f35ca\" (UID: \"44e80226-1db6-4df7-9561-eb33be2f35ca\") " Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.069276 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44e80226-1db6-4df7-9561-eb33be2f35ca-inventory\") pod \"44e80226-1db6-4df7-9561-eb33be2f35ca\" (UID: \"44e80226-1db6-4df7-9561-eb33be2f35ca\") " Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.076667 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44e80226-1db6-4df7-9561-eb33be2f35ca-kube-api-access-7ttdf" (OuterVolumeSpecName: "kube-api-access-7ttdf") pod "44e80226-1db6-4df7-9561-eb33be2f35ca" (UID: "44e80226-1db6-4df7-9561-eb33be2f35ca"). InnerVolumeSpecName "kube-api-access-7ttdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.106792 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44e80226-1db6-4df7-9561-eb33be2f35ca-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "44e80226-1db6-4df7-9561-eb33be2f35ca" (UID: "44e80226-1db6-4df7-9561-eb33be2f35ca"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.107784 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44e80226-1db6-4df7-9561-eb33be2f35ca-inventory" (OuterVolumeSpecName: "inventory") pod "44e80226-1db6-4df7-9561-eb33be2f35ca" (UID: "44e80226-1db6-4df7-9561-eb33be2f35ca"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.172041 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/44e80226-1db6-4df7-9561-eb33be2f35ca-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.172083 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7ttdf\" (UniqueName: \"kubernetes.io/projected/44e80226-1db6-4df7-9561-eb33be2f35ca-kube-api-access-7ttdf\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.172099 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/44e80226-1db6-4df7-9561-eb33be2f35ca-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.362625 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nzdh" event={"ID":"2914b248-f67d-49a8-b908-8fed6028f477","Type":"ContainerStarted","Data":"a077a1adb2d5dd376f1d4ed3c1ee47c5e9273d4607826b898271a6592f910c89"} Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.365862 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.365826 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-wflfz" event={"ID":"44e80226-1db6-4df7-9561-eb33be2f35ca","Type":"ContainerDied","Data":"9d77fd63df0c020a4a7d5a12a58917d8e312d3e3e853c47aef01638d0b53a551"} Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.365983 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9d77fd63df0c020a4a7d5a12a58917d8e312d3e3e853c47aef01638d0b53a551" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.446910 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p"] Mar 13 09:47:41 crc kubenswrapper[4930]: E0313 09:47:41.447646 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44e80226-1db6-4df7-9561-eb33be2f35ca" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.447725 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="44e80226-1db6-4df7-9561-eb33be2f35ca" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.448065 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="44e80226-1db6-4df7-9561-eb33be2f35ca" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.448934 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.452210 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.452635 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.452716 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.453599 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.467880 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p"] Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.582099 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fspj\" (UniqueName: \"kubernetes.io/projected/0e34f769-6a2b-4088-b913-6f5cdb7721a8-kube-api-access-6fspj\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p\" (UID: \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.582551 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e34f769-6a2b-4088-b913-6f5cdb7721a8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p\" (UID: \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.582615 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0e34f769-6a2b-4088-b913-6f5cdb7721a8-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p\" (UID: \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.684917 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e34f769-6a2b-4088-b913-6f5cdb7721a8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p\" (UID: \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.684990 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0e34f769-6a2b-4088-b913-6f5cdb7721a8-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p\" (UID: \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.685162 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fspj\" (UniqueName: \"kubernetes.io/projected/0e34f769-6a2b-4088-b913-6f5cdb7721a8-kube-api-access-6fspj\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p\" (UID: \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.689769 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0e34f769-6a2b-4088-b913-6f5cdb7721a8-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p\" (UID: \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.689849 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e34f769-6a2b-4088-b913-6f5cdb7721a8-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p\" (UID: \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.701110 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fspj\" (UniqueName: \"kubernetes.io/projected/0e34f769-6a2b-4088-b913-6f5cdb7721a8-kube-api-access-6fspj\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p\" (UID: \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:47:41 crc kubenswrapper[4930]: I0313 09:47:41.768508 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:47:42 crc kubenswrapper[4930]: I0313 09:47:42.050717 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-3a0f-account-create-update-82l86"] Mar 13 09:47:42 crc kubenswrapper[4930]: I0313 09:47:42.067683 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-r7nvk"] Mar 13 09:47:42 crc kubenswrapper[4930]: I0313 09:47:42.078162 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-3a0f-account-create-update-82l86"] Mar 13 09:47:42 crc kubenswrapper[4930]: I0313 09:47:42.088139 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-r7nvk"] Mar 13 09:47:42 crc kubenswrapper[4930]: I0313 09:47:42.357463 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p"] Mar 13 09:47:42 crc kubenswrapper[4930]: W0313 09:47:42.365994 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0e34f769_6a2b_4088_b913_6f5cdb7721a8.slice/crio-33ec364f0732858f5af9f0330de2011b4f93fe6738514f11376aed689eb92549 WatchSource:0}: Error finding container 33ec364f0732858f5af9f0330de2011b4f93fe6738514f11376aed689eb92549: Status 404 returned error can't find the container with id 33ec364f0732858f5af9f0330de2011b4f93fe6738514f11376aed689eb92549 Mar 13 09:47:42 crc kubenswrapper[4930]: I0313 09:47:42.383674 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" event={"ID":"0e34f769-6a2b-4088-b913-6f5cdb7721a8","Type":"ContainerStarted","Data":"33ec364f0732858f5af9f0330de2011b4f93fe6738514f11376aed689eb92549"} Mar 13 09:47:43 crc kubenswrapper[4930]: I0313 09:47:43.394427 4930 generic.go:334] "Generic (PLEG): container finished" podID="2914b248-f67d-49a8-b908-8fed6028f477" containerID="a077a1adb2d5dd376f1d4ed3c1ee47c5e9273d4607826b898271a6592f910c89" exitCode=0 Mar 13 09:47:43 crc kubenswrapper[4930]: I0313 09:47:43.394473 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nzdh" event={"ID":"2914b248-f67d-49a8-b908-8fed6028f477","Type":"ContainerDied","Data":"a077a1adb2d5dd376f1d4ed3c1ee47c5e9273d4607826b898271a6592f910c89"} Mar 13 09:47:43 crc kubenswrapper[4930]: I0313 09:47:43.405288 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" event={"ID":"0e34f769-6a2b-4088-b913-6f5cdb7721a8","Type":"ContainerStarted","Data":"a10c18940a584061a0b5439a420706c87c30b5f15bc574ec83eed6eb9a25ba7a"} Mar 13 09:47:43 crc kubenswrapper[4930]: I0313 09:47:43.998488 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05b2d756-3fdc-4914-b511-350ee81fabb4" path="/var/lib/kubelet/pods/05b2d756-3fdc-4914-b511-350ee81fabb4/volumes" Mar 13 09:47:44 crc kubenswrapper[4930]: I0313 09:47:44.005607 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5c00fcf-72f0-45e6-a1b3-235677ec4d14" path="/var/lib/kubelet/pods/e5c00fcf-72f0-45e6-a1b3-235677ec4d14/volumes" Mar 13 09:47:45 crc kubenswrapper[4930]: I0313 09:47:45.427370 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nzdh" event={"ID":"2914b248-f67d-49a8-b908-8fed6028f477","Type":"ContainerStarted","Data":"0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053"} Mar 13 09:47:45 crc kubenswrapper[4930]: I0313 09:47:45.445527 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" podStartSLOduration=3.925693333 podStartE2EDuration="4.445507054s" podCreationTimestamp="2026-03-13 09:47:41 +0000 UTC" firstStartedPulling="2026-03-13 09:47:42.368562355 +0000 UTC m=+2103.118477032" lastFinishedPulling="2026-03-13 09:47:42.888376076 +0000 UTC m=+2103.638290753" observedRunningTime="2026-03-13 09:47:43.435333929 +0000 UTC m=+2104.185248606" watchObservedRunningTime="2026-03-13 09:47:45.445507054 +0000 UTC m=+2106.195421731" Mar 13 09:47:45 crc kubenswrapper[4930]: I0313 09:47:45.450580 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7nzdh" podStartSLOduration=2.336161174 podStartE2EDuration="6.45056622s" podCreationTimestamp="2026-03-13 09:47:39 +0000 UTC" firstStartedPulling="2026-03-13 09:47:40.342789202 +0000 UTC m=+2101.092703879" lastFinishedPulling="2026-03-13 09:47:44.457194248 +0000 UTC m=+2105.207108925" observedRunningTime="2026-03-13 09:47:45.444388227 +0000 UTC m=+2106.194302904" watchObservedRunningTime="2026-03-13 09:47:45.45056622 +0000 UTC m=+2106.200480897" Mar 13 09:47:49 crc kubenswrapper[4930]: I0313 09:47:49.027162 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-hljq9"] Mar 13 09:47:49 crc kubenswrapper[4930]: I0313 09:47:49.059555 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-hljq9"] Mar 13 09:47:49 crc kubenswrapper[4930]: I0313 09:47:49.374283 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:49 crc kubenswrapper[4930]: I0313 09:47:49.374638 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:49 crc kubenswrapper[4930]: I0313 09:47:49.441354 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:49 crc kubenswrapper[4930]: I0313 09:47:49.517371 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:49 crc kubenswrapper[4930]: I0313 09:47:49.681149 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7nzdh"] Mar 13 09:47:49 crc kubenswrapper[4930]: I0313 09:47:49.988574 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67ceb05f-14a6-4adb-af90-5d52549e4f23" path="/var/lib/kubelet/pods/67ceb05f-14a6-4adb-af90-5d52549e4f23/volumes" Mar 13 09:47:51 crc kubenswrapper[4930]: I0313 09:47:51.488383 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7nzdh" podUID="2914b248-f67d-49a8-b908-8fed6028f477" containerName="registry-server" containerID="cri-o://0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053" gracePeriod=2 Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.466869 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.470804 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pd6j\" (UniqueName: \"kubernetes.io/projected/2914b248-f67d-49a8-b908-8fed6028f477-kube-api-access-8pd6j\") pod \"2914b248-f67d-49a8-b908-8fed6028f477\" (UID: \"2914b248-f67d-49a8-b908-8fed6028f477\") " Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.470876 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2914b248-f67d-49a8-b908-8fed6028f477-utilities\") pod \"2914b248-f67d-49a8-b908-8fed6028f477\" (UID: \"2914b248-f67d-49a8-b908-8fed6028f477\") " Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.470906 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2914b248-f67d-49a8-b908-8fed6028f477-catalog-content\") pod \"2914b248-f67d-49a8-b908-8fed6028f477\" (UID: \"2914b248-f67d-49a8-b908-8fed6028f477\") " Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.471801 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2914b248-f67d-49a8-b908-8fed6028f477-utilities" (OuterVolumeSpecName: "utilities") pod "2914b248-f67d-49a8-b908-8fed6028f477" (UID: "2914b248-f67d-49a8-b908-8fed6028f477"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.472301 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2914b248-f67d-49a8-b908-8fed6028f477-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.476627 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2914b248-f67d-49a8-b908-8fed6028f477-kube-api-access-8pd6j" (OuterVolumeSpecName: "kube-api-access-8pd6j") pod "2914b248-f67d-49a8-b908-8fed6028f477" (UID: "2914b248-f67d-49a8-b908-8fed6028f477"). InnerVolumeSpecName "kube-api-access-8pd6j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.514767 4930 generic.go:334] "Generic (PLEG): container finished" podID="2914b248-f67d-49a8-b908-8fed6028f477" containerID="0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053" exitCode=0 Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.514814 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nzdh" event={"ID":"2914b248-f67d-49a8-b908-8fed6028f477","Type":"ContainerDied","Data":"0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053"} Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.514841 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7nzdh" event={"ID":"2914b248-f67d-49a8-b908-8fed6028f477","Type":"ContainerDied","Data":"39f64ceac38f2d1c4f4b92038f243c6417c46be970ff20e094e310d10ab55a2e"} Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.514868 4930 scope.go:117] "RemoveContainer" containerID="0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.515023 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7nzdh" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.568211 4930 scope.go:117] "RemoveContainer" containerID="a077a1adb2d5dd376f1d4ed3c1ee47c5e9273d4607826b898271a6592f910c89" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.575013 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pd6j\" (UniqueName: \"kubernetes.io/projected/2914b248-f67d-49a8-b908-8fed6028f477-kube-api-access-8pd6j\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.595537 4930 scope.go:117] "RemoveContainer" containerID="4f1dd9a3b8991c95b91d81a584b59d04a6e71ee611d88a5cb38e1451872b14f2" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.649348 4930 scope.go:117] "RemoveContainer" containerID="0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053" Mar 13 09:47:52 crc kubenswrapper[4930]: E0313 09:47:52.649789 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053\": container with ID starting with 0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053 not found: ID does not exist" containerID="0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.649817 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053"} err="failed to get container status \"0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053\": rpc error: code = NotFound desc = could not find container \"0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053\": container with ID starting with 0fbce7612a178bca8ba51e4424e505b99fe175026bde17f1bc4f756337459053 not found: ID does not exist" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.649836 4930 scope.go:117] "RemoveContainer" containerID="a077a1adb2d5dd376f1d4ed3c1ee47c5e9273d4607826b898271a6592f910c89" Mar 13 09:47:52 crc kubenswrapper[4930]: E0313 09:47:52.650086 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a077a1adb2d5dd376f1d4ed3c1ee47c5e9273d4607826b898271a6592f910c89\": container with ID starting with a077a1adb2d5dd376f1d4ed3c1ee47c5e9273d4607826b898271a6592f910c89 not found: ID does not exist" containerID="a077a1adb2d5dd376f1d4ed3c1ee47c5e9273d4607826b898271a6592f910c89" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.650113 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a077a1adb2d5dd376f1d4ed3c1ee47c5e9273d4607826b898271a6592f910c89"} err="failed to get container status \"a077a1adb2d5dd376f1d4ed3c1ee47c5e9273d4607826b898271a6592f910c89\": rpc error: code = NotFound desc = could not find container \"a077a1adb2d5dd376f1d4ed3c1ee47c5e9273d4607826b898271a6592f910c89\": container with ID starting with a077a1adb2d5dd376f1d4ed3c1ee47c5e9273d4607826b898271a6592f910c89 not found: ID does not exist" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.650128 4930 scope.go:117] "RemoveContainer" containerID="4f1dd9a3b8991c95b91d81a584b59d04a6e71ee611d88a5cb38e1451872b14f2" Mar 13 09:47:52 crc kubenswrapper[4930]: E0313 09:47:52.650402 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f1dd9a3b8991c95b91d81a584b59d04a6e71ee611d88a5cb38e1451872b14f2\": container with ID starting with 4f1dd9a3b8991c95b91d81a584b59d04a6e71ee611d88a5cb38e1451872b14f2 not found: ID does not exist" containerID="4f1dd9a3b8991c95b91d81a584b59d04a6e71ee611d88a5cb38e1451872b14f2" Mar 13 09:47:52 crc kubenswrapper[4930]: I0313 09:47:52.650521 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f1dd9a3b8991c95b91d81a584b59d04a6e71ee611d88a5cb38e1451872b14f2"} err="failed to get container status \"4f1dd9a3b8991c95b91d81a584b59d04a6e71ee611d88a5cb38e1451872b14f2\": rpc error: code = NotFound desc = could not find container \"4f1dd9a3b8991c95b91d81a584b59d04a6e71ee611d88a5cb38e1451872b14f2\": container with ID starting with 4f1dd9a3b8991c95b91d81a584b59d04a6e71ee611d88a5cb38e1451872b14f2 not found: ID does not exist" Mar 13 09:47:53 crc kubenswrapper[4930]: I0313 09:47:53.562200 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2914b248-f67d-49a8-b908-8fed6028f477-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2914b248-f67d-49a8-b908-8fed6028f477" (UID: "2914b248-f67d-49a8-b908-8fed6028f477"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:47:53 crc kubenswrapper[4930]: I0313 09:47:53.606510 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2914b248-f67d-49a8-b908-8fed6028f477-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:47:53 crc kubenswrapper[4930]: I0313 09:47:53.749934 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7nzdh"] Mar 13 09:47:53 crc kubenswrapper[4930]: I0313 09:47:53.760253 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7nzdh"] Mar 13 09:47:53 crc kubenswrapper[4930]: I0313 09:47:53.983456 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2914b248-f67d-49a8-b908-8fed6028f477" path="/var/lib/kubelet/pods/2914b248-f67d-49a8-b908-8fed6028f477/volumes" Mar 13 09:47:54 crc kubenswrapper[4930]: I0313 09:47:54.042022 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8pbts"] Mar 13 09:47:54 crc kubenswrapper[4930]: I0313 09:47:54.056315 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-8pbts"] Mar 13 09:47:56 crc kubenswrapper[4930]: I0313 09:47:56.003294 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7ccda94d-acb1-4c49-91e1-db24243cf6a9" path="/var/lib/kubelet/pods/7ccda94d-acb1-4c49-91e1-db24243cf6a9/volumes" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.146287 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556588-l9kj2"] Mar 13 09:48:00 crc kubenswrapper[4930]: E0313 09:48:00.148450 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2914b248-f67d-49a8-b908-8fed6028f477" containerName="extract-content" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.148474 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2914b248-f67d-49a8-b908-8fed6028f477" containerName="extract-content" Mar 13 09:48:00 crc kubenswrapper[4930]: E0313 09:48:00.148784 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2914b248-f67d-49a8-b908-8fed6028f477" containerName="extract-utilities" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.148793 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2914b248-f67d-49a8-b908-8fed6028f477" containerName="extract-utilities" Mar 13 09:48:00 crc kubenswrapper[4930]: E0313 09:48:00.149052 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2914b248-f67d-49a8-b908-8fed6028f477" containerName="registry-server" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.149133 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2914b248-f67d-49a8-b908-8fed6028f477" containerName="registry-server" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.149992 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="2914b248-f67d-49a8-b908-8fed6028f477" containerName="registry-server" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.154635 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556588-l9kj2" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.157177 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.157320 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.158286 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.158466 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556588-l9kj2"] Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.254482 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrw62\" (UniqueName: \"kubernetes.io/projected/08e202d3-7e4e-40e8-aa35-c3579e1b8f2d-kube-api-access-xrw62\") pod \"auto-csr-approver-29556588-l9kj2\" (UID: \"08e202d3-7e4e-40e8-aa35-c3579e1b8f2d\") " pod="openshift-infra/auto-csr-approver-29556588-l9kj2" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.357482 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrw62\" (UniqueName: \"kubernetes.io/projected/08e202d3-7e4e-40e8-aa35-c3579e1b8f2d-kube-api-access-xrw62\") pod \"auto-csr-approver-29556588-l9kj2\" (UID: \"08e202d3-7e4e-40e8-aa35-c3579e1b8f2d\") " pod="openshift-infra/auto-csr-approver-29556588-l9kj2" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.379679 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrw62\" (UniqueName: \"kubernetes.io/projected/08e202d3-7e4e-40e8-aa35-c3579e1b8f2d-kube-api-access-xrw62\") pod \"auto-csr-approver-29556588-l9kj2\" (UID: \"08e202d3-7e4e-40e8-aa35-c3579e1b8f2d\") " pod="openshift-infra/auto-csr-approver-29556588-l9kj2" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.487871 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556588-l9kj2" Mar 13 09:48:00 crc kubenswrapper[4930]: I0313 09:48:00.932156 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556588-l9kj2"] Mar 13 09:48:01 crc kubenswrapper[4930]: I0313 09:48:01.616004 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556588-l9kj2" event={"ID":"08e202d3-7e4e-40e8-aa35-c3579e1b8f2d","Type":"ContainerStarted","Data":"842350f1f70ffd47caf5db481b368fc2dafffc0cbfa7e767c173e48e70e3546c"} Mar 13 09:48:02 crc kubenswrapper[4930]: I0313 09:48:02.629063 4930 generic.go:334] "Generic (PLEG): container finished" podID="08e202d3-7e4e-40e8-aa35-c3579e1b8f2d" containerID="b570600fe74e7d4c56540dda5e350fd1e7efd08ca6e5e1b86c8e8315c9f83c9e" exitCode=0 Mar 13 09:48:02 crc kubenswrapper[4930]: I0313 09:48:02.629102 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556588-l9kj2" event={"ID":"08e202d3-7e4e-40e8-aa35-c3579e1b8f2d","Type":"ContainerDied","Data":"b570600fe74e7d4c56540dda5e350fd1e7efd08ca6e5e1b86c8e8315c9f83c9e"} Mar 13 09:48:04 crc kubenswrapper[4930]: I0313 09:48:04.074562 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556588-l9kj2" Mar 13 09:48:04 crc kubenswrapper[4930]: I0313 09:48:04.239274 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrw62\" (UniqueName: \"kubernetes.io/projected/08e202d3-7e4e-40e8-aa35-c3579e1b8f2d-kube-api-access-xrw62\") pod \"08e202d3-7e4e-40e8-aa35-c3579e1b8f2d\" (UID: \"08e202d3-7e4e-40e8-aa35-c3579e1b8f2d\") " Mar 13 09:48:04 crc kubenswrapper[4930]: I0313 09:48:04.245532 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08e202d3-7e4e-40e8-aa35-c3579e1b8f2d-kube-api-access-xrw62" (OuterVolumeSpecName: "kube-api-access-xrw62") pod "08e202d3-7e4e-40e8-aa35-c3579e1b8f2d" (UID: "08e202d3-7e4e-40e8-aa35-c3579e1b8f2d"). InnerVolumeSpecName "kube-api-access-xrw62". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:48:04 crc kubenswrapper[4930]: I0313 09:48:04.342739 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrw62\" (UniqueName: \"kubernetes.io/projected/08e202d3-7e4e-40e8-aa35-c3579e1b8f2d-kube-api-access-xrw62\") on node \"crc\" DevicePath \"\"" Mar 13 09:48:04 crc kubenswrapper[4930]: I0313 09:48:04.655080 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556588-l9kj2" event={"ID":"08e202d3-7e4e-40e8-aa35-c3579e1b8f2d","Type":"ContainerDied","Data":"842350f1f70ffd47caf5db481b368fc2dafffc0cbfa7e767c173e48e70e3546c"} Mar 13 09:48:04 crc kubenswrapper[4930]: I0313 09:48:04.655128 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="842350f1f70ffd47caf5db481b368fc2dafffc0cbfa7e767c173e48e70e3546c" Mar 13 09:48:04 crc kubenswrapper[4930]: I0313 09:48:04.655159 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556588-l9kj2" Mar 13 09:48:05 crc kubenswrapper[4930]: I0313 09:48:05.179187 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556582-g29pq"] Mar 13 09:48:05 crc kubenswrapper[4930]: I0313 09:48:05.192865 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556582-g29pq"] Mar 13 09:48:05 crc kubenswrapper[4930]: I0313 09:48:05.985812 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c927b774-514d-483a-97ec-d72cbc62a7a3" path="/var/lib/kubelet/pods/c927b774-514d-483a-97ec-d72cbc62a7a3/volumes" Mar 13 09:48:28 crc kubenswrapper[4930]: I0313 09:48:28.928198 4930 generic.go:334] "Generic (PLEG): container finished" podID="0e34f769-6a2b-4088-b913-6f5cdb7721a8" containerID="a10c18940a584061a0b5439a420706c87c30b5f15bc574ec83eed6eb9a25ba7a" exitCode=0 Mar 13 09:48:28 crc kubenswrapper[4930]: I0313 09:48:28.928270 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" event={"ID":"0e34f769-6a2b-4088-b913-6f5cdb7721a8","Type":"ContainerDied","Data":"a10c18940a584061a0b5439a420706c87c30b5f15bc574ec83eed6eb9a25ba7a"} Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.217164 4930 scope.go:117] "RemoveContainer" containerID="7c2d12ac50c8e12bd82ecd79c65662aa39860ec7c3c0512ad6f6716b34b3987d" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.248422 4930 scope.go:117] "RemoveContainer" containerID="3adf722854cc2bf589fb56a6a42c8ea43ad76ee7cc5596cb8c0a20c02739b258" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.332085 4930 scope.go:117] "RemoveContainer" containerID="e7f08920cc72e929f43e7095a0eee6181dd89a73297ca086d458bf87dfa2aa88" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.378277 4930 scope.go:117] "RemoveContainer" containerID="e55d97b040c4f9dd7afc52670411587db48d28ec123410d9c0cd8f8198000fc7" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.494744 4930 scope.go:117] "RemoveContainer" containerID="3b32b5ccd512cccbb3e04056bfad8de1f17b714917b17aac986da63abd41b2f0" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.543157 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.733717 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fspj\" (UniqueName: \"kubernetes.io/projected/0e34f769-6a2b-4088-b913-6f5cdb7721a8-kube-api-access-6fspj\") pod \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\" (UID: \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\") " Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.734836 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0e34f769-6a2b-4088-b913-6f5cdb7721a8-ssh-key-openstack-edpm-ipam\") pod \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\" (UID: \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\") " Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.734901 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e34f769-6a2b-4088-b913-6f5cdb7721a8-inventory\") pod \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\" (UID: \"0e34f769-6a2b-4088-b913-6f5cdb7721a8\") " Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.741600 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e34f769-6a2b-4088-b913-6f5cdb7721a8-kube-api-access-6fspj" (OuterVolumeSpecName: "kube-api-access-6fspj") pod "0e34f769-6a2b-4088-b913-6f5cdb7721a8" (UID: "0e34f769-6a2b-4088-b913-6f5cdb7721a8"). InnerVolumeSpecName "kube-api-access-6fspj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.771833 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e34f769-6a2b-4088-b913-6f5cdb7721a8-inventory" (OuterVolumeSpecName: "inventory") pod "0e34f769-6a2b-4088-b913-6f5cdb7721a8" (UID: "0e34f769-6a2b-4088-b913-6f5cdb7721a8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.775056 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e34f769-6a2b-4088-b913-6f5cdb7721a8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0e34f769-6a2b-4088-b913-6f5cdb7721a8" (UID: "0e34f769-6a2b-4088-b913-6f5cdb7721a8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.838772 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fspj\" (UniqueName: \"kubernetes.io/projected/0e34f769-6a2b-4088-b913-6f5cdb7721a8-kube-api-access-6fspj\") on node \"crc\" DevicePath \"\"" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.839026 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0e34f769-6a2b-4088-b913-6f5cdb7721a8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.839036 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0e34f769-6a2b-4088-b913-6f5cdb7721a8-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.947867 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" event={"ID":"0e34f769-6a2b-4088-b913-6f5cdb7721a8","Type":"ContainerDied","Data":"33ec364f0732858f5af9f0330de2011b4f93fe6738514f11376aed689eb92549"} Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.947919 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33ec364f0732858f5af9f0330de2011b4f93fe6738514f11376aed689eb92549" Mar 13 09:48:30 crc kubenswrapper[4930]: I0313 09:48:30.947934 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vpb8p" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.109696 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-xt8z2"] Mar 13 09:48:31 crc kubenswrapper[4930]: E0313 09:48:31.110328 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08e202d3-7e4e-40e8-aa35-c3579e1b8f2d" containerName="oc" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.110351 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="08e202d3-7e4e-40e8-aa35-c3579e1b8f2d" containerName="oc" Mar 13 09:48:31 crc kubenswrapper[4930]: E0313 09:48:31.110372 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e34f769-6a2b-4088-b913-6f5cdb7721a8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.110382 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e34f769-6a2b-4088-b913-6f5cdb7721a8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.110693 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="08e202d3-7e4e-40e8-aa35-c3579e1b8f2d" containerName="oc" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.110745 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e34f769-6a2b-4088-b913-6f5cdb7721a8" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.111826 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.114700 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.114892 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.116201 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.116275 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.127486 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-xt8z2"] Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.250497 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-xt8z2\" (UID: \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\") " pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.250546 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-xt8z2\" (UID: \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\") " pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.250714 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2cqn\" (UniqueName: \"kubernetes.io/projected/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-kube-api-access-t2cqn\") pod \"ssh-known-hosts-edpm-deployment-xt8z2\" (UID: \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\") " pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.353350 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-xt8z2\" (UID: \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\") " pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.353408 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-xt8z2\" (UID: \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\") " pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.353477 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2cqn\" (UniqueName: \"kubernetes.io/projected/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-kube-api-access-t2cqn\") pod \"ssh-known-hosts-edpm-deployment-xt8z2\" (UID: \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\") " pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.358712 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-xt8z2\" (UID: \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\") " pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.361387 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-xt8z2\" (UID: \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\") " pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.375923 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2cqn\" (UniqueName: \"kubernetes.io/projected/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-kube-api-access-t2cqn\") pod \"ssh-known-hosts-edpm-deployment-xt8z2\" (UID: \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\") " pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:31 crc kubenswrapper[4930]: I0313 09:48:31.439533 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:32 crc kubenswrapper[4930]: I0313 09:48:32.045604 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-xt8z2"] Mar 13 09:48:32 crc kubenswrapper[4930]: I0313 09:48:32.981515 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" event={"ID":"3dd69158-ad05-4ae2-95dc-e228e75f9b2f","Type":"ContainerStarted","Data":"5824e5470707774688d560ec691ce3225aaaa556418f166cd3a57bf32d64ad9e"} Mar 13 09:48:32 crc kubenswrapper[4930]: I0313 09:48:32.982139 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" event={"ID":"3dd69158-ad05-4ae2-95dc-e228e75f9b2f","Type":"ContainerStarted","Data":"64bfd567c8a703934a12997dbc377a1a2faecd8191cd9e362fccb72172481727"} Mar 13 09:48:33 crc kubenswrapper[4930]: I0313 09:48:33.003337 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" podStartSLOduration=1.466717278 podStartE2EDuration="2.003319383s" podCreationTimestamp="2026-03-13 09:48:31 +0000 UTC" firstStartedPulling="2026-03-13 09:48:32.050115673 +0000 UTC m=+2152.800030350" lastFinishedPulling="2026-03-13 09:48:32.586717778 +0000 UTC m=+2153.336632455" observedRunningTime="2026-03-13 09:48:32.997645141 +0000 UTC m=+2153.747559838" watchObservedRunningTime="2026-03-13 09:48:33.003319383 +0000 UTC m=+2153.753234060" Mar 13 09:48:36 crc kubenswrapper[4930]: I0313 09:48:36.047617 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-h78lz"] Mar 13 09:48:36 crc kubenswrapper[4930]: I0313 09:48:36.058160 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-h78lz"] Mar 13 09:48:37 crc kubenswrapper[4930]: I0313 09:48:37.984286 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b218ee1b-e381-4f43-b72b-8b11b440448f" path="/var/lib/kubelet/pods/b218ee1b-e381-4f43-b72b-8b11b440448f/volumes" Mar 13 09:48:40 crc kubenswrapper[4930]: I0313 09:48:40.050746 4930 generic.go:334] "Generic (PLEG): container finished" podID="3dd69158-ad05-4ae2-95dc-e228e75f9b2f" containerID="5824e5470707774688d560ec691ce3225aaaa556418f166cd3a57bf32d64ad9e" exitCode=0 Mar 13 09:48:40 crc kubenswrapper[4930]: I0313 09:48:40.050823 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" event={"ID":"3dd69158-ad05-4ae2-95dc-e228e75f9b2f","Type":"ContainerDied","Data":"5824e5470707774688d560ec691ce3225aaaa556418f166cd3a57bf32d64ad9e"} Mar 13 09:48:41 crc kubenswrapper[4930]: I0313 09:48:41.559415 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:41 crc kubenswrapper[4930]: I0313 09:48:41.702718 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-inventory-0\") pod \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\" (UID: \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\") " Mar 13 09:48:41 crc kubenswrapper[4930]: I0313 09:48:41.702886 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-ssh-key-openstack-edpm-ipam\") pod \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\" (UID: \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\") " Mar 13 09:48:41 crc kubenswrapper[4930]: I0313 09:48:41.703316 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2cqn\" (UniqueName: \"kubernetes.io/projected/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-kube-api-access-t2cqn\") pod \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\" (UID: \"3dd69158-ad05-4ae2-95dc-e228e75f9b2f\") " Mar 13 09:48:41 crc kubenswrapper[4930]: I0313 09:48:41.727762 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-kube-api-access-t2cqn" (OuterVolumeSpecName: "kube-api-access-t2cqn") pod "3dd69158-ad05-4ae2-95dc-e228e75f9b2f" (UID: "3dd69158-ad05-4ae2-95dc-e228e75f9b2f"). InnerVolumeSpecName "kube-api-access-t2cqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:48:41 crc kubenswrapper[4930]: I0313 09:48:41.762834 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "3dd69158-ad05-4ae2-95dc-e228e75f9b2f" (UID: "3dd69158-ad05-4ae2-95dc-e228e75f9b2f"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:48:41 crc kubenswrapper[4930]: I0313 09:48:41.806073 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2cqn\" (UniqueName: \"kubernetes.io/projected/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-kube-api-access-t2cqn\") on node \"crc\" DevicePath \"\"" Mar 13 09:48:41 crc kubenswrapper[4930]: I0313 09:48:41.806111 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:48:41 crc kubenswrapper[4930]: I0313 09:48:41.850549 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "3dd69158-ad05-4ae2-95dc-e228e75f9b2f" (UID: "3dd69158-ad05-4ae2-95dc-e228e75f9b2f"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:48:41 crc kubenswrapper[4930]: I0313 09:48:41.908261 4930 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/3dd69158-ad05-4ae2-95dc-e228e75f9b2f-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.074489 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" event={"ID":"3dd69158-ad05-4ae2-95dc-e228e75f9b2f","Type":"ContainerDied","Data":"64bfd567c8a703934a12997dbc377a1a2faecd8191cd9e362fccb72172481727"} Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.074732 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64bfd567c8a703934a12997dbc377a1a2faecd8191cd9e362fccb72172481727" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.074556 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-xt8z2" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.142279 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm"] Mar 13 09:48:42 crc kubenswrapper[4930]: E0313 09:48:42.143017 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd69158-ad05-4ae2-95dc-e228e75f9b2f" containerName="ssh-known-hosts-edpm-deployment" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.143097 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd69158-ad05-4ae2-95dc-e228e75f9b2f" containerName="ssh-known-hosts-edpm-deployment" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.143365 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dd69158-ad05-4ae2-95dc-e228e75f9b2f" containerName="ssh-known-hosts-edpm-deployment" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.144237 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.146215 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.146396 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.147507 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.149378 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.161166 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm"] Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.216327 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/825008ba-b6ee-4824-831c-5bfce06c02d8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-b2ggm\" (UID: \"825008ba-b6ee-4824-831c-5bfce06c02d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.216419 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqdgj\" (UniqueName: \"kubernetes.io/projected/825008ba-b6ee-4824-831c-5bfce06c02d8-kube-api-access-kqdgj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-b2ggm\" (UID: \"825008ba-b6ee-4824-831c-5bfce06c02d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.216485 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/825008ba-b6ee-4824-831c-5bfce06c02d8-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-b2ggm\" (UID: \"825008ba-b6ee-4824-831c-5bfce06c02d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.318689 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/825008ba-b6ee-4824-831c-5bfce06c02d8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-b2ggm\" (UID: \"825008ba-b6ee-4824-831c-5bfce06c02d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.318783 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqdgj\" (UniqueName: \"kubernetes.io/projected/825008ba-b6ee-4824-831c-5bfce06c02d8-kube-api-access-kqdgj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-b2ggm\" (UID: \"825008ba-b6ee-4824-831c-5bfce06c02d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.318835 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/825008ba-b6ee-4824-831c-5bfce06c02d8-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-b2ggm\" (UID: \"825008ba-b6ee-4824-831c-5bfce06c02d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.323134 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/825008ba-b6ee-4824-831c-5bfce06c02d8-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-b2ggm\" (UID: \"825008ba-b6ee-4824-831c-5bfce06c02d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.323593 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/825008ba-b6ee-4824-831c-5bfce06c02d8-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-b2ggm\" (UID: \"825008ba-b6ee-4824-831c-5bfce06c02d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.336713 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqdgj\" (UniqueName: \"kubernetes.io/projected/825008ba-b6ee-4824-831c-5bfce06c02d8-kube-api-access-kqdgj\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-b2ggm\" (UID: \"825008ba-b6ee-4824-831c-5bfce06c02d8\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:42 crc kubenswrapper[4930]: I0313 09:48:42.486395 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:43 crc kubenswrapper[4930]: I0313 09:48:43.015009 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm"] Mar 13 09:48:43 crc kubenswrapper[4930]: I0313 09:48:43.088077 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" event={"ID":"825008ba-b6ee-4824-831c-5bfce06c02d8","Type":"ContainerStarted","Data":"275e209bacf6737b588631277aa05e8ef8770dab3c8841c3dbc0ad261b8fd308"} Mar 13 09:48:44 crc kubenswrapper[4930]: I0313 09:48:44.101232 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" event={"ID":"825008ba-b6ee-4824-831c-5bfce06c02d8","Type":"ContainerStarted","Data":"52bb4c0ba43e86e46243a44c70d4f05080dbcc885dfcf2114d344af3ba115bf7"} Mar 13 09:48:44 crc kubenswrapper[4930]: I0313 09:48:44.127846 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" podStartSLOduration=1.6900635510000002 podStartE2EDuration="2.127819315s" podCreationTimestamp="2026-03-13 09:48:42 +0000 UTC" firstStartedPulling="2026-03-13 09:48:43.032940224 +0000 UTC m=+2163.782854901" lastFinishedPulling="2026-03-13 09:48:43.470695988 +0000 UTC m=+2164.220610665" observedRunningTime="2026-03-13 09:48:44.116853361 +0000 UTC m=+2164.866768038" watchObservedRunningTime="2026-03-13 09:48:44.127819315 +0000 UTC m=+2164.877733992" Mar 13 09:48:51 crc kubenswrapper[4930]: I0313 09:48:51.165690 4930 generic.go:334] "Generic (PLEG): container finished" podID="825008ba-b6ee-4824-831c-5bfce06c02d8" containerID="52bb4c0ba43e86e46243a44c70d4f05080dbcc885dfcf2114d344af3ba115bf7" exitCode=0 Mar 13 09:48:51 crc kubenswrapper[4930]: I0313 09:48:51.165769 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" event={"ID":"825008ba-b6ee-4824-831c-5bfce06c02d8","Type":"ContainerDied","Data":"52bb4c0ba43e86e46243a44c70d4f05080dbcc885dfcf2114d344af3ba115bf7"} Mar 13 09:48:52 crc kubenswrapper[4930]: I0313 09:48:52.802062 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:52 crc kubenswrapper[4930]: I0313 09:48:52.878960 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/825008ba-b6ee-4824-831c-5bfce06c02d8-ssh-key-openstack-edpm-ipam\") pod \"825008ba-b6ee-4824-831c-5bfce06c02d8\" (UID: \"825008ba-b6ee-4824-831c-5bfce06c02d8\") " Mar 13 09:48:52 crc kubenswrapper[4930]: I0313 09:48:52.879240 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/825008ba-b6ee-4824-831c-5bfce06c02d8-inventory\") pod \"825008ba-b6ee-4824-831c-5bfce06c02d8\" (UID: \"825008ba-b6ee-4824-831c-5bfce06c02d8\") " Mar 13 09:48:52 crc kubenswrapper[4930]: I0313 09:48:52.879271 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqdgj\" (UniqueName: \"kubernetes.io/projected/825008ba-b6ee-4824-831c-5bfce06c02d8-kube-api-access-kqdgj\") pod \"825008ba-b6ee-4824-831c-5bfce06c02d8\" (UID: \"825008ba-b6ee-4824-831c-5bfce06c02d8\") " Mar 13 09:48:52 crc kubenswrapper[4930]: I0313 09:48:52.886848 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/825008ba-b6ee-4824-831c-5bfce06c02d8-kube-api-access-kqdgj" (OuterVolumeSpecName: "kube-api-access-kqdgj") pod "825008ba-b6ee-4824-831c-5bfce06c02d8" (UID: "825008ba-b6ee-4824-831c-5bfce06c02d8"). InnerVolumeSpecName "kube-api-access-kqdgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:48:52 crc kubenswrapper[4930]: I0313 09:48:52.939707 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825008ba-b6ee-4824-831c-5bfce06c02d8-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "825008ba-b6ee-4824-831c-5bfce06c02d8" (UID: "825008ba-b6ee-4824-831c-5bfce06c02d8"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:48:52 crc kubenswrapper[4930]: I0313 09:48:52.945783 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/825008ba-b6ee-4824-831c-5bfce06c02d8-inventory" (OuterVolumeSpecName: "inventory") pod "825008ba-b6ee-4824-831c-5bfce06c02d8" (UID: "825008ba-b6ee-4824-831c-5bfce06c02d8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:48:52 crc kubenswrapper[4930]: I0313 09:48:52.981565 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/825008ba-b6ee-4824-831c-5bfce06c02d8-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:48:52 crc kubenswrapper[4930]: I0313 09:48:52.981596 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqdgj\" (UniqueName: \"kubernetes.io/projected/825008ba-b6ee-4824-831c-5bfce06c02d8-kube-api-access-kqdgj\") on node \"crc\" DevicePath \"\"" Mar 13 09:48:52 crc kubenswrapper[4930]: I0313 09:48:52.981610 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/825008ba-b6ee-4824-831c-5bfce06c02d8-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.184915 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" event={"ID":"825008ba-b6ee-4824-831c-5bfce06c02d8","Type":"ContainerDied","Data":"275e209bacf6737b588631277aa05e8ef8770dab3c8841c3dbc0ad261b8fd308"} Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.184954 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="275e209bacf6737b588631277aa05e8ef8770dab3c8841c3dbc0ad261b8fd308" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.185000 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-b2ggm" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.270740 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5"] Mar 13 09:48:53 crc kubenswrapper[4930]: E0313 09:48:53.271520 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="825008ba-b6ee-4824-831c-5bfce06c02d8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.271684 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="825008ba-b6ee-4824-831c-5bfce06c02d8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.272083 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="825008ba-b6ee-4824-831c-5bfce06c02d8" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.273181 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.275370 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.275411 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.275450 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.276349 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.287660 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5"] Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.389846 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m828n\" (UniqueName: \"kubernetes.io/projected/cf6c3054-4042-4a44-b5ea-fd9903657425-kube-api-access-m828n\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5\" (UID: \"cf6c3054-4042-4a44-b5ea-fd9903657425\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.390223 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf6c3054-4042-4a44-b5ea-fd9903657425-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5\" (UID: \"cf6c3054-4042-4a44-b5ea-fd9903657425\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.390306 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf6c3054-4042-4a44-b5ea-fd9903657425-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5\" (UID: \"cf6c3054-4042-4a44-b5ea-fd9903657425\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.492664 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf6c3054-4042-4a44-b5ea-fd9903657425-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5\" (UID: \"cf6c3054-4042-4a44-b5ea-fd9903657425\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.492793 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf6c3054-4042-4a44-b5ea-fd9903657425-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5\" (UID: \"cf6c3054-4042-4a44-b5ea-fd9903657425\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.492841 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m828n\" (UniqueName: \"kubernetes.io/projected/cf6c3054-4042-4a44-b5ea-fd9903657425-kube-api-access-m828n\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5\" (UID: \"cf6c3054-4042-4a44-b5ea-fd9903657425\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.496679 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf6c3054-4042-4a44-b5ea-fd9903657425-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5\" (UID: \"cf6c3054-4042-4a44-b5ea-fd9903657425\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.500984 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf6c3054-4042-4a44-b5ea-fd9903657425-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5\" (UID: \"cf6c3054-4042-4a44-b5ea-fd9903657425\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.509054 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m828n\" (UniqueName: \"kubernetes.io/projected/cf6c3054-4042-4a44-b5ea-fd9903657425-kube-api-access-m828n\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5\" (UID: \"cf6c3054-4042-4a44-b5ea-fd9903657425\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:48:53 crc kubenswrapper[4930]: I0313 09:48:53.603536 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:48:54 crc kubenswrapper[4930]: I0313 09:48:54.179713 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5"] Mar 13 09:48:54 crc kubenswrapper[4930]: I0313 09:48:54.195661 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" event={"ID":"cf6c3054-4042-4a44-b5ea-fd9903657425","Type":"ContainerStarted","Data":"d99d5b7050d1e3af3999a31f2086967f47814a1e5101772ab67b951d81de7e20"} Mar 13 09:48:55 crc kubenswrapper[4930]: I0313 09:48:55.209010 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" event={"ID":"cf6c3054-4042-4a44-b5ea-fd9903657425","Type":"ContainerStarted","Data":"4376b9b8942dca19d7a439fffca8d709cb81107eb3bc1efc041eac2a4ee01034"} Mar 13 09:48:55 crc kubenswrapper[4930]: I0313 09:48:55.227481 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" podStartSLOduration=1.523421916 podStartE2EDuration="2.227465346s" podCreationTimestamp="2026-03-13 09:48:53 +0000 UTC" firstStartedPulling="2026-03-13 09:48:54.166201706 +0000 UTC m=+2174.916116383" lastFinishedPulling="2026-03-13 09:48:54.870245146 +0000 UTC m=+2175.620159813" observedRunningTime="2026-03-13 09:48:55.22442568 +0000 UTC m=+2175.974340347" watchObservedRunningTime="2026-03-13 09:48:55.227465346 +0000 UTC m=+2175.977380013" Mar 13 09:49:04 crc kubenswrapper[4930]: I0313 09:49:04.307965 4930 generic.go:334] "Generic (PLEG): container finished" podID="cf6c3054-4042-4a44-b5ea-fd9903657425" containerID="4376b9b8942dca19d7a439fffca8d709cb81107eb3bc1efc041eac2a4ee01034" exitCode=0 Mar 13 09:49:04 crc kubenswrapper[4930]: I0313 09:49:04.308591 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" event={"ID":"cf6c3054-4042-4a44-b5ea-fd9903657425","Type":"ContainerDied","Data":"4376b9b8942dca19d7a439fffca8d709cb81107eb3bc1efc041eac2a4ee01034"} Mar 13 09:49:05 crc kubenswrapper[4930]: I0313 09:49:05.847482 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:49:05 crc kubenswrapper[4930]: I0313 09:49:05.926468 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m828n\" (UniqueName: \"kubernetes.io/projected/cf6c3054-4042-4a44-b5ea-fd9903657425-kube-api-access-m828n\") pod \"cf6c3054-4042-4a44-b5ea-fd9903657425\" (UID: \"cf6c3054-4042-4a44-b5ea-fd9903657425\") " Mar 13 09:49:05 crc kubenswrapper[4930]: I0313 09:49:05.926954 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf6c3054-4042-4a44-b5ea-fd9903657425-ssh-key-openstack-edpm-ipam\") pod \"cf6c3054-4042-4a44-b5ea-fd9903657425\" (UID: \"cf6c3054-4042-4a44-b5ea-fd9903657425\") " Mar 13 09:49:05 crc kubenswrapper[4930]: I0313 09:49:05.927136 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf6c3054-4042-4a44-b5ea-fd9903657425-inventory\") pod \"cf6c3054-4042-4a44-b5ea-fd9903657425\" (UID: \"cf6c3054-4042-4a44-b5ea-fd9903657425\") " Mar 13 09:49:05 crc kubenswrapper[4930]: I0313 09:49:05.931929 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cf6c3054-4042-4a44-b5ea-fd9903657425-kube-api-access-m828n" (OuterVolumeSpecName: "kube-api-access-m828n") pod "cf6c3054-4042-4a44-b5ea-fd9903657425" (UID: "cf6c3054-4042-4a44-b5ea-fd9903657425"). InnerVolumeSpecName "kube-api-access-m828n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:49:05 crc kubenswrapper[4930]: I0313 09:49:05.961259 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf6c3054-4042-4a44-b5ea-fd9903657425-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "cf6c3054-4042-4a44-b5ea-fd9903657425" (UID: "cf6c3054-4042-4a44-b5ea-fd9903657425"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:49:05 crc kubenswrapper[4930]: I0313 09:49:05.962992 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cf6c3054-4042-4a44-b5ea-fd9903657425-inventory" (OuterVolumeSpecName: "inventory") pod "cf6c3054-4042-4a44-b5ea-fd9903657425" (UID: "cf6c3054-4042-4a44-b5ea-fd9903657425"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.030031 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m828n\" (UniqueName: \"kubernetes.io/projected/cf6c3054-4042-4a44-b5ea-fd9903657425-kube-api-access-m828n\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.030065 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cf6c3054-4042-4a44-b5ea-fd9903657425-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.030081 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cf6c3054-4042-4a44-b5ea-fd9903657425-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.329721 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" event={"ID":"cf6c3054-4042-4a44-b5ea-fd9903657425","Type":"ContainerDied","Data":"d99d5b7050d1e3af3999a31f2086967f47814a1e5101772ab67b951d81de7e20"} Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.330246 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d99d5b7050d1e3af3999a31f2086967f47814a1e5101772ab67b951d81de7e20" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.329769 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zbff5" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.412918 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh"] Mar 13 09:49:06 crc kubenswrapper[4930]: E0313 09:49:06.414758 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cf6c3054-4042-4a44-b5ea-fd9903657425" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.414779 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="cf6c3054-4042-4a44-b5ea-fd9903657425" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.416579 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="cf6c3054-4042-4a44-b5ea-fd9903657425" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.417417 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.420558 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.420641 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.420562 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.421032 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.421121 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.421178 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.421250 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.422579 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.422795 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.425955 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh"] Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.541692 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.541746 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.541786 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.541983 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.542344 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.542397 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.542494 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.542544 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.542572 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.542672 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.542880 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.542991 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.543035 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.543105 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b6kf\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-kube-api-access-8b6kf\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.543235 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.543326 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646340 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646383 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646417 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646473 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646493 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646523 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646570 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646596 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646631 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646654 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b6kf\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-kube-api-access-8b6kf\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646686 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646706 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646759 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646800 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646830 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.646853 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.653291 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.653532 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.653762 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.653876 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.654287 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.654588 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.654693 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.654784 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.655280 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.655645 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.655763 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.655791 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.656099 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.656100 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.664378 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b6kf\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-kube-api-access-8b6kf\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.664883 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:06 crc kubenswrapper[4930]: I0313 09:49:06.743370 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:07 crc kubenswrapper[4930]: I0313 09:49:07.291835 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh"] Mar 13 09:49:07 crc kubenswrapper[4930]: I0313 09:49:07.339928 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" event={"ID":"edad7c64-c7cc-4c8c-9c6c-ba74236085c6","Type":"ContainerStarted","Data":"672531285f8724a4f0579a5d2c3988002a2e3d5dc3453f1c245a47593968d9bc"} Mar 13 09:49:08 crc kubenswrapper[4930]: I0313 09:49:08.353552 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" event={"ID":"edad7c64-c7cc-4c8c-9c6c-ba74236085c6","Type":"ContainerStarted","Data":"976dbf19e2b0c3f5fd8a1fc9a985919ea24ed90237035929b9dccefbb134a463"} Mar 13 09:49:12 crc kubenswrapper[4930]: I0313 09:49:12.307818 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:49:12 crc kubenswrapper[4930]: I0313 09:49:12.308331 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:49:30 crc kubenswrapper[4930]: I0313 09:49:30.691134 4930 scope.go:117] "RemoveContainer" containerID="c616c3f491277e469695830551caa9d391677e210c54ae9ecd436ba968511443" Mar 13 09:49:42 crc kubenswrapper[4930]: I0313 09:49:42.308954 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:49:42 crc kubenswrapper[4930]: I0313 09:49:42.310384 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:49:47 crc kubenswrapper[4930]: I0313 09:49:47.767780 4930 generic.go:334] "Generic (PLEG): container finished" podID="edad7c64-c7cc-4c8c-9c6c-ba74236085c6" containerID="976dbf19e2b0c3f5fd8a1fc9a985919ea24ed90237035929b9dccefbb134a463" exitCode=0 Mar 13 09:49:47 crc kubenswrapper[4930]: I0313 09:49:47.767888 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" event={"ID":"edad7c64-c7cc-4c8c-9c6c-ba74236085c6","Type":"ContainerDied","Data":"976dbf19e2b0c3f5fd8a1fc9a985919ea24ed90237035929b9dccefbb134a463"} Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.299260 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361189 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361235 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361301 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-ovn-combined-ca-bundle\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361353 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-ssh-key-openstack-edpm-ipam\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361400 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-libvirt-combined-ca-bundle\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361515 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-repo-setup-combined-ca-bundle\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361532 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-bootstrap-combined-ca-bundle\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361576 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361609 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-neutron-metadata-combined-ca-bundle\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361659 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-nova-combined-ca-bundle\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361710 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-ovn-default-certs-0\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361752 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-inventory\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361800 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361847 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b6kf\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-kube-api-access-8b6kf\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361884 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-telemetry-power-monitoring-combined-ca-bundle\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.361906 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-telemetry-combined-ca-bundle\") pod \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\" (UID: \"edad7c64-c7cc-4c8c-9c6c-ba74236085c6\") " Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.367635 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.367774 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.369845 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.369974 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.371060 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.371131 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.371616 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.377071 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.377656 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.383636 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.384622 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-kube-api-access-8b6kf" (OuterVolumeSpecName: "kube-api-access-8b6kf") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "kube-api-access-8b6kf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.384997 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.387587 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.393136 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.402226 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.428275 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-inventory" (OuterVolumeSpecName: "inventory") pod "edad7c64-c7cc-4c8c-9c6c-ba74236085c6" (UID: "edad7c64-c7cc-4c8c-9c6c-ba74236085c6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.464860 4930 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.464905 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b6kf\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-kube-api-access-8b6kf\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.464916 4930 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.464926 4930 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.464939 4930 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.464949 4930 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.464960 4930 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.464972 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.464980 4930 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.464989 4930 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.465000 4930 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.465009 4930 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.465018 4930 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.465028 4930 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.465038 4930 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.465046 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/edad7c64-c7cc-4c8c-9c6c-ba74236085c6-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.801408 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" event={"ID":"edad7c64-c7cc-4c8c-9c6c-ba74236085c6","Type":"ContainerDied","Data":"672531285f8724a4f0579a5d2c3988002a2e3d5dc3453f1c245a47593968d9bc"} Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.801473 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="672531285f8724a4f0579a5d2c3988002a2e3d5dc3453f1c245a47593968d9bc" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.801947 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-2s9mh" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.901547 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx"] Mar 13 09:49:49 crc kubenswrapper[4930]: E0313 09:49:49.902134 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="edad7c64-c7cc-4c8c-9c6c-ba74236085c6" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.902160 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="edad7c64-c7cc-4c8c-9c6c-ba74236085c6" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.902465 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="edad7c64-c7cc-4c8c-9c6c-ba74236085c6" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.903574 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.906891 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.907497 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.907845 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.908095 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.909020 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.915263 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx"] Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.975172 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/adf80189-23e8-4e0c-83c5-c961aa84e191-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.975457 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cwfz\" (UniqueName: \"kubernetes.io/projected/adf80189-23e8-4e0c-83c5-c961aa84e191-kube-api-access-6cwfz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.975612 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.975775 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:49 crc kubenswrapper[4930]: I0313 09:49:49.975826 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:50 crc kubenswrapper[4930]: I0313 09:49:50.077677 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/adf80189-23e8-4e0c-83c5-c961aa84e191-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:50 crc kubenswrapper[4930]: I0313 09:49:50.077727 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cwfz\" (UniqueName: \"kubernetes.io/projected/adf80189-23e8-4e0c-83c5-c961aa84e191-kube-api-access-6cwfz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:50 crc kubenswrapper[4930]: I0313 09:49:50.077763 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:50 crc kubenswrapper[4930]: I0313 09:49:50.077818 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:50 crc kubenswrapper[4930]: I0313 09:49:50.077899 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:50 crc kubenswrapper[4930]: I0313 09:49:50.078768 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/adf80189-23e8-4e0c-83c5-c961aa84e191-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:50 crc kubenswrapper[4930]: I0313 09:49:50.082127 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:50 crc kubenswrapper[4930]: I0313 09:49:50.082405 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:50 crc kubenswrapper[4930]: I0313 09:49:50.083665 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:50 crc kubenswrapper[4930]: I0313 09:49:50.098324 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cwfz\" (UniqueName: \"kubernetes.io/projected/adf80189-23e8-4e0c-83c5-c961aa84e191-kube-api-access-6cwfz\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-pxwlx\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:51 crc kubenswrapper[4930]: I0313 09:49:51.174306 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:49:51 crc kubenswrapper[4930]: I0313 09:49:51.815245 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx"] Mar 13 09:49:52 crc kubenswrapper[4930]: I0313 09:49:52.215929 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" event={"ID":"adf80189-23e8-4e0c-83c5-c961aa84e191","Type":"ContainerStarted","Data":"1363908dd506f59691e9dff800497656ae7c582d65e8a7575ebc9c6093b12612"} Mar 13 09:49:53 crc kubenswrapper[4930]: I0313 09:49:53.225753 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" event={"ID":"adf80189-23e8-4e0c-83c5-c961aa84e191","Type":"ContainerStarted","Data":"038396c47c6e8efbe1d87a47538c4d3652fac01ee42ddc32a6df1b375a8c72cb"} Mar 13 09:49:53 crc kubenswrapper[4930]: I0313 09:49:53.244024 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" podStartSLOduration=3.8369865880000003 podStartE2EDuration="4.244002473s" podCreationTimestamp="2026-03-13 09:49:49 +0000 UTC" firstStartedPulling="2026-03-13 09:49:51.818517768 +0000 UTC m=+2232.568432445" lastFinishedPulling="2026-03-13 09:49:52.225533653 +0000 UTC m=+2232.975448330" observedRunningTime="2026-03-13 09:49:53.240483745 +0000 UTC m=+2233.990398422" watchObservedRunningTime="2026-03-13 09:49:53.244002473 +0000 UTC m=+2233.993917150" Mar 13 09:50:00 crc kubenswrapper[4930]: I0313 09:50:00.139779 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556590-hnzrg"] Mar 13 09:50:00 crc kubenswrapper[4930]: I0313 09:50:00.142937 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556590-hnzrg" Mar 13 09:50:00 crc kubenswrapper[4930]: I0313 09:50:00.145845 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:50:00 crc kubenswrapper[4930]: I0313 09:50:00.148824 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:50:00 crc kubenswrapper[4930]: I0313 09:50:00.148900 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:50:00 crc kubenswrapper[4930]: I0313 09:50:00.151309 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556590-hnzrg"] Mar 13 09:50:00 crc kubenswrapper[4930]: I0313 09:50:00.218975 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjdvl\" (UniqueName: \"kubernetes.io/projected/abd4c5da-c943-4cd3-9138-7b5c0cc39aef-kube-api-access-xjdvl\") pod \"auto-csr-approver-29556590-hnzrg\" (UID: \"abd4c5da-c943-4cd3-9138-7b5c0cc39aef\") " pod="openshift-infra/auto-csr-approver-29556590-hnzrg" Mar 13 09:50:00 crc kubenswrapper[4930]: I0313 09:50:00.322259 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjdvl\" (UniqueName: \"kubernetes.io/projected/abd4c5da-c943-4cd3-9138-7b5c0cc39aef-kube-api-access-xjdvl\") pod \"auto-csr-approver-29556590-hnzrg\" (UID: \"abd4c5da-c943-4cd3-9138-7b5c0cc39aef\") " pod="openshift-infra/auto-csr-approver-29556590-hnzrg" Mar 13 09:50:00 crc kubenswrapper[4930]: I0313 09:50:00.363141 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjdvl\" (UniqueName: \"kubernetes.io/projected/abd4c5da-c943-4cd3-9138-7b5c0cc39aef-kube-api-access-xjdvl\") pod \"auto-csr-approver-29556590-hnzrg\" (UID: \"abd4c5da-c943-4cd3-9138-7b5c0cc39aef\") " pod="openshift-infra/auto-csr-approver-29556590-hnzrg" Mar 13 09:50:00 crc kubenswrapper[4930]: I0313 09:50:00.469190 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556590-hnzrg" Mar 13 09:50:00 crc kubenswrapper[4930]: I0313 09:50:00.930171 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556590-hnzrg"] Mar 13 09:50:01 crc kubenswrapper[4930]: I0313 09:50:01.311920 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556590-hnzrg" event={"ID":"abd4c5da-c943-4cd3-9138-7b5c0cc39aef","Type":"ContainerStarted","Data":"c79da57044fe15be04d061bc0beebc4f5c1a6c4f9461f1a8f3ca1a852ef23222"} Mar 13 09:50:03 crc kubenswrapper[4930]: I0313 09:50:03.331510 4930 generic.go:334] "Generic (PLEG): container finished" podID="abd4c5da-c943-4cd3-9138-7b5c0cc39aef" containerID="6e93b7d86ada97ba1bd5dfa206bcdca731cac44253657583d74c90eeaf1244d9" exitCode=0 Mar 13 09:50:03 crc kubenswrapper[4930]: I0313 09:50:03.331805 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556590-hnzrg" event={"ID":"abd4c5da-c943-4cd3-9138-7b5c0cc39aef","Type":"ContainerDied","Data":"6e93b7d86ada97ba1bd5dfa206bcdca731cac44253657583d74c90eeaf1244d9"} Mar 13 09:50:04 crc kubenswrapper[4930]: I0313 09:50:04.744343 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556590-hnzrg" Mar 13 09:50:04 crc kubenswrapper[4930]: I0313 09:50:04.779888 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjdvl\" (UniqueName: \"kubernetes.io/projected/abd4c5da-c943-4cd3-9138-7b5c0cc39aef-kube-api-access-xjdvl\") pod \"abd4c5da-c943-4cd3-9138-7b5c0cc39aef\" (UID: \"abd4c5da-c943-4cd3-9138-7b5c0cc39aef\") " Mar 13 09:50:04 crc kubenswrapper[4930]: I0313 09:50:04.785224 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/abd4c5da-c943-4cd3-9138-7b5c0cc39aef-kube-api-access-xjdvl" (OuterVolumeSpecName: "kube-api-access-xjdvl") pod "abd4c5da-c943-4cd3-9138-7b5c0cc39aef" (UID: "abd4c5da-c943-4cd3-9138-7b5c0cc39aef"). InnerVolumeSpecName "kube-api-access-xjdvl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:50:04 crc kubenswrapper[4930]: I0313 09:50:04.883298 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjdvl\" (UniqueName: \"kubernetes.io/projected/abd4c5da-c943-4cd3-9138-7b5c0cc39aef-kube-api-access-xjdvl\") on node \"crc\" DevicePath \"\"" Mar 13 09:50:05 crc kubenswrapper[4930]: I0313 09:50:05.355855 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556590-hnzrg" event={"ID":"abd4c5da-c943-4cd3-9138-7b5c0cc39aef","Type":"ContainerDied","Data":"c79da57044fe15be04d061bc0beebc4f5c1a6c4f9461f1a8f3ca1a852ef23222"} Mar 13 09:50:05 crc kubenswrapper[4930]: I0313 09:50:05.355899 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c79da57044fe15be04d061bc0beebc4f5c1a6c4f9461f1a8f3ca1a852ef23222" Mar 13 09:50:05 crc kubenswrapper[4930]: I0313 09:50:05.355937 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556590-hnzrg" Mar 13 09:50:05 crc kubenswrapper[4930]: I0313 09:50:05.824053 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556584-qmpnl"] Mar 13 09:50:05 crc kubenswrapper[4930]: I0313 09:50:05.834298 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556584-qmpnl"] Mar 13 09:50:05 crc kubenswrapper[4930]: I0313 09:50:05.986166 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ef9b7b54-f041-4a1f-84bf-f8b2ee52818d" path="/var/lib/kubelet/pods/ef9b7b54-f041-4a1f-84bf-f8b2ee52818d/volumes" Mar 13 09:50:12 crc kubenswrapper[4930]: I0313 09:50:12.309002 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:50:12 crc kubenswrapper[4930]: I0313 09:50:12.309705 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:50:12 crc kubenswrapper[4930]: I0313 09:50:12.309766 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:50:12 crc kubenswrapper[4930]: I0313 09:50:12.310788 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 09:50:12 crc kubenswrapper[4930]: I0313 09:50:12.310892 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" gracePeriod=600 Mar 13 09:50:12 crc kubenswrapper[4930]: E0313 09:50:12.430720 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:50:13 crc kubenswrapper[4930]: I0313 09:50:13.444658 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" exitCode=0 Mar 13 09:50:13 crc kubenswrapper[4930]: I0313 09:50:13.444726 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f"} Mar 13 09:50:13 crc kubenswrapper[4930]: I0313 09:50:13.445105 4930 scope.go:117] "RemoveContainer" containerID="9d9e501a7dfec9be2a4a7aae030dcc1a1a33117fc6ee6f36debdcb82b8cfee89" Mar 13 09:50:13 crc kubenswrapper[4930]: I0313 09:50:13.445922 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:50:13 crc kubenswrapper[4930]: E0313 09:50:13.446312 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:50:19 crc kubenswrapper[4930]: I0313 09:50:19.038375 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-6sjqb"] Mar 13 09:50:19 crc kubenswrapper[4930]: I0313 09:50:19.060201 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-6sjqb"] Mar 13 09:50:19 crc kubenswrapper[4930]: I0313 09:50:19.987202 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94958e65-6d03-4346-b6c0-47e3de576961" path="/var/lib/kubelet/pods/94958e65-6d03-4346-b6c0-47e3de576961/volumes" Mar 13 09:50:26 crc kubenswrapper[4930]: I0313 09:50:26.971703 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:50:26 crc kubenswrapper[4930]: E0313 09:50:26.972539 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:50:30 crc kubenswrapper[4930]: I0313 09:50:30.767490 4930 scope.go:117] "RemoveContainer" containerID="7fcf530a30accc88731658ebc0a7c9693a939ead6904aeeaeefeda5b176ec3dc" Mar 13 09:50:30 crc kubenswrapper[4930]: I0313 09:50:30.813344 4930 scope.go:117] "RemoveContainer" containerID="73fd01a431aee4e9e432af3440c1d2dd1c2c77f8d2ac43502cac74a3c95ceff6" Mar 13 09:50:40 crc kubenswrapper[4930]: I0313 09:50:40.971788 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:50:40 crc kubenswrapper[4930]: E0313 09:50:40.972604 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:50:49 crc kubenswrapper[4930]: I0313 09:50:49.809351 4930 generic.go:334] "Generic (PLEG): container finished" podID="adf80189-23e8-4e0c-83c5-c961aa84e191" containerID="038396c47c6e8efbe1d87a47538c4d3652fac01ee42ddc32a6df1b375a8c72cb" exitCode=0 Mar 13 09:50:49 crc kubenswrapper[4930]: I0313 09:50:49.809513 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" event={"ID":"adf80189-23e8-4e0c-83c5-c961aa84e191","Type":"ContainerDied","Data":"038396c47c6e8efbe1d87a47538c4d3652fac01ee42ddc32a6df1b375a8c72cb"} Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.314857 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.497349 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-ovn-combined-ca-bundle\") pod \"adf80189-23e8-4e0c-83c5-c961aa84e191\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.498511 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cwfz\" (UniqueName: \"kubernetes.io/projected/adf80189-23e8-4e0c-83c5-c961aa84e191-kube-api-access-6cwfz\") pod \"adf80189-23e8-4e0c-83c5-c961aa84e191\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.498776 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/adf80189-23e8-4e0c-83c5-c961aa84e191-ovncontroller-config-0\") pod \"adf80189-23e8-4e0c-83c5-c961aa84e191\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.498980 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-inventory\") pod \"adf80189-23e8-4e0c-83c5-c961aa84e191\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.499082 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-ssh-key-openstack-edpm-ipam\") pod \"adf80189-23e8-4e0c-83c5-c961aa84e191\" (UID: \"adf80189-23e8-4e0c-83c5-c961aa84e191\") " Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.517646 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "adf80189-23e8-4e0c-83c5-c961aa84e191" (UID: "adf80189-23e8-4e0c-83c5-c961aa84e191"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.517677 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adf80189-23e8-4e0c-83c5-c961aa84e191-kube-api-access-6cwfz" (OuterVolumeSpecName: "kube-api-access-6cwfz") pod "adf80189-23e8-4e0c-83c5-c961aa84e191" (UID: "adf80189-23e8-4e0c-83c5-c961aa84e191"). InnerVolumeSpecName "kube-api-access-6cwfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.532951 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "adf80189-23e8-4e0c-83c5-c961aa84e191" (UID: "adf80189-23e8-4e0c-83c5-c961aa84e191"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.538920 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-inventory" (OuterVolumeSpecName: "inventory") pod "adf80189-23e8-4e0c-83c5-c961aa84e191" (UID: "adf80189-23e8-4e0c-83c5-c961aa84e191"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.548607 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adf80189-23e8-4e0c-83c5-c961aa84e191-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "adf80189-23e8-4e0c-83c5-c961aa84e191" (UID: "adf80189-23e8-4e0c-83c5-c961aa84e191"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.604329 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.604379 4930 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.604392 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cwfz\" (UniqueName: \"kubernetes.io/projected/adf80189-23e8-4e0c-83c5-c961aa84e191-kube-api-access-6cwfz\") on node \"crc\" DevicePath \"\"" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.604404 4930 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/adf80189-23e8-4e0c-83c5-c961aa84e191-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.604419 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/adf80189-23e8-4e0c-83c5-c961aa84e191-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.829996 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" event={"ID":"adf80189-23e8-4e0c-83c5-c961aa84e191","Type":"ContainerDied","Data":"1363908dd506f59691e9dff800497656ae7c582d65e8a7575ebc9c6093b12612"} Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.830051 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1363908dd506f59691e9dff800497656ae7c582d65e8a7575ebc9c6093b12612" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.830046 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-pxwlx" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.923235 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj"] Mar 13 09:50:51 crc kubenswrapper[4930]: E0313 09:50:51.923716 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="abd4c5da-c943-4cd3-9138-7b5c0cc39aef" containerName="oc" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.923731 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="abd4c5da-c943-4cd3-9138-7b5c0cc39aef" containerName="oc" Mar 13 09:50:51 crc kubenswrapper[4930]: E0313 09:50:51.923774 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adf80189-23e8-4e0c-83c5-c961aa84e191" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.923783 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="adf80189-23e8-4e0c-83c5-c961aa84e191" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.924022 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="adf80189-23e8-4e0c-83c5-c961aa84e191" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.924061 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="abd4c5da-c943-4cd3-9138-7b5c0cc39aef" containerName="oc" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.924839 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.927792 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.928003 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.929744 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.929749 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.929787 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.930117 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:50:51 crc kubenswrapper[4930]: I0313 09:50:51.937085 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj"] Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.014823 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.014909 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.015019 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.015131 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.015353 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-44zqp\" (UniqueName: \"kubernetes.io/projected/9fd633a8-da0c-4f2d-ae82-ca7205d75103-kube-api-access-44zqp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.015470 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.117593 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.117656 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.117694 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.117746 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.117820 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-44zqp\" (UniqueName: \"kubernetes.io/projected/9fd633a8-da0c-4f2d-ae82-ca7205d75103-kube-api-access-44zqp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.117853 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.122383 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.122512 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.122621 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.123097 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.124150 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.135944 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-44zqp\" (UniqueName: \"kubernetes.io/projected/9fd633a8-da0c-4f2d-ae82-ca7205d75103-kube-api-access-44zqp\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.260045 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.805544 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj"] Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.825562 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 09:50:52 crc kubenswrapper[4930]: I0313 09:50:52.844327 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" event={"ID":"9fd633a8-da0c-4f2d-ae82-ca7205d75103","Type":"ContainerStarted","Data":"23b192cb3fa4607bb103007f3a3e57d8fd271ca769445fa7bc177abadcafbef6"} Mar 13 09:50:53 crc kubenswrapper[4930]: I0313 09:50:53.855017 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" event={"ID":"9fd633a8-da0c-4f2d-ae82-ca7205d75103","Type":"ContainerStarted","Data":"4314c5eb1eed0e841875d7f6e845911c769d8d018f5673e683865a8e5937b625"} Mar 13 09:50:53 crc kubenswrapper[4930]: I0313 09:50:53.880956 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" podStartSLOduration=2.393599378 podStartE2EDuration="2.880938081s" podCreationTimestamp="2026-03-13 09:50:51 +0000 UTC" firstStartedPulling="2026-03-13 09:50:52.825353473 +0000 UTC m=+2293.575268140" lastFinishedPulling="2026-03-13 09:50:53.312692166 +0000 UTC m=+2294.062606843" observedRunningTime="2026-03-13 09:50:53.873387382 +0000 UTC m=+2294.623302069" watchObservedRunningTime="2026-03-13 09:50:53.880938081 +0000 UTC m=+2294.630852758" Mar 13 09:50:54 crc kubenswrapper[4930]: I0313 09:50:54.971281 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:50:54 crc kubenswrapper[4930]: E0313 09:50:54.971969 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:50:57 crc kubenswrapper[4930]: I0313 09:50:57.047974 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-4rcdw"] Mar 13 09:50:57 crc kubenswrapper[4930]: I0313 09:50:57.059175 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-4rcdw"] Mar 13 09:50:57 crc kubenswrapper[4930]: I0313 09:50:57.985362 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18da2b4c-e41d-4e0c-b470-b9651e7f1aae" path="/var/lib/kubelet/pods/18da2b4c-e41d-4e0c-b470-b9651e7f1aae/volumes" Mar 13 09:51:07 crc kubenswrapper[4930]: I0313 09:51:07.971480 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:51:07 crc kubenswrapper[4930]: E0313 09:51:07.972198 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:51:21 crc kubenswrapper[4930]: I0313 09:51:21.971866 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:51:21 crc kubenswrapper[4930]: E0313 09:51:21.972763 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:51:30 crc kubenswrapper[4930]: I0313 09:51:30.904987 4930 scope.go:117] "RemoveContainer" containerID="65b572c3008282bf84784b2e4acabdb2a2c4f34701ed1143700ac0cdaca63253" Mar 13 09:51:34 crc kubenswrapper[4930]: I0313 09:51:34.970621 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:51:34 crc kubenswrapper[4930]: E0313 09:51:34.971128 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:51:39 crc kubenswrapper[4930]: I0313 09:51:39.335273 4930 generic.go:334] "Generic (PLEG): container finished" podID="9fd633a8-da0c-4f2d-ae82-ca7205d75103" containerID="4314c5eb1eed0e841875d7f6e845911c769d8d018f5673e683865a8e5937b625" exitCode=0 Mar 13 09:51:39 crc kubenswrapper[4930]: I0313 09:51:39.335352 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" event={"ID":"9fd633a8-da0c-4f2d-ae82-ca7205d75103","Type":"ContainerDied","Data":"4314c5eb1eed0e841875d7f6e845911c769d8d018f5673e683865a8e5937b625"} Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.825850 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.920973 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-neutron-ovn-metadata-agent-neutron-config-0\") pod \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.921097 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-ssh-key-openstack-edpm-ipam\") pod \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.921139 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-neutron-metadata-combined-ca-bundle\") pod \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.921254 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-44zqp\" (UniqueName: \"kubernetes.io/projected/9fd633a8-da0c-4f2d-ae82-ca7205d75103-kube-api-access-44zqp\") pod \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.921307 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-inventory\") pod \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.921334 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-nova-metadata-neutron-config-0\") pod \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\" (UID: \"9fd633a8-da0c-4f2d-ae82-ca7205d75103\") " Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.927064 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fd633a8-da0c-4f2d-ae82-ca7205d75103-kube-api-access-44zqp" (OuterVolumeSpecName: "kube-api-access-44zqp") pod "9fd633a8-da0c-4f2d-ae82-ca7205d75103" (UID: "9fd633a8-da0c-4f2d-ae82-ca7205d75103"). InnerVolumeSpecName "kube-api-access-44zqp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.927907 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "9fd633a8-da0c-4f2d-ae82-ca7205d75103" (UID: "9fd633a8-da0c-4f2d-ae82-ca7205d75103"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.954957 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9fd633a8-da0c-4f2d-ae82-ca7205d75103" (UID: "9fd633a8-da0c-4f2d-ae82-ca7205d75103"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.958422 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "9fd633a8-da0c-4f2d-ae82-ca7205d75103" (UID: "9fd633a8-da0c-4f2d-ae82-ca7205d75103"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.959254 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "9fd633a8-da0c-4f2d-ae82-ca7205d75103" (UID: "9fd633a8-da0c-4f2d-ae82-ca7205d75103"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:51:40 crc kubenswrapper[4930]: I0313 09:51:40.981231 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-inventory" (OuterVolumeSpecName: "inventory") pod "9fd633a8-da0c-4f2d-ae82-ca7205d75103" (UID: "9fd633a8-da0c-4f2d-ae82-ca7205d75103"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.025215 4930 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.025255 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.025267 4930 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.025278 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-44zqp\" (UniqueName: \"kubernetes.io/projected/9fd633a8-da0c-4f2d-ae82-ca7205d75103-kube-api-access-44zqp\") on node \"crc\" DevicePath \"\"" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.025289 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.025298 4930 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/9fd633a8-da0c-4f2d-ae82-ca7205d75103-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.365104 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" event={"ID":"9fd633a8-da0c-4f2d-ae82-ca7205d75103","Type":"ContainerDied","Data":"23b192cb3fa4607bb103007f3a3e57d8fd271ca769445fa7bc177abadcafbef6"} Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.365138 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-wzkbj" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.365162 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23b192cb3fa4607bb103007f3a3e57d8fd271ca769445fa7bc177abadcafbef6" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.471354 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf"] Mar 13 09:51:41 crc kubenswrapper[4930]: E0313 09:51:41.472055 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fd633a8-da0c-4f2d-ae82-ca7205d75103" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.472073 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fd633a8-da0c-4f2d-ae82-ca7205d75103" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.472309 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fd633a8-da0c-4f2d-ae82-ca7205d75103" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.473396 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.476859 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.481334 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.481407 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.481487 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.481757 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.512284 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf"] Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.541199 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.541531 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.541699 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.541833 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d448r\" (UniqueName: \"kubernetes.io/projected/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-kube-api-access-d448r\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.542250 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.645245 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.645291 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d448r\" (UniqueName: \"kubernetes.io/projected/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-kube-api-access-d448r\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.645369 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.645481 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.645509 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.649400 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.649903 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.649998 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.650066 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.668219 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d448r\" (UniqueName: \"kubernetes.io/projected/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-kube-api-access-d448r\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-6sghf\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:41 crc kubenswrapper[4930]: I0313 09:51:41.797835 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:51:42 crc kubenswrapper[4930]: I0313 09:51:42.375054 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf"] Mar 13 09:51:43 crc kubenswrapper[4930]: I0313 09:51:43.394135 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" event={"ID":"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1","Type":"ContainerStarted","Data":"df63cf641d5abb8fc04f8ac933f6d5119fbbaeb369daa43ee46a4ffee07137c5"} Mar 13 09:51:43 crc kubenswrapper[4930]: I0313 09:51:43.394743 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" event={"ID":"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1","Type":"ContainerStarted","Data":"5fd0541ec729d7afea54f67407ce0cb0a0f4692803f9e6c7fea93eb00e13ad78"} Mar 13 09:51:43 crc kubenswrapper[4930]: I0313 09:51:43.419727 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" podStartSLOduration=1.787916021 podStartE2EDuration="2.419704085s" podCreationTimestamp="2026-03-13 09:51:41 +0000 UTC" firstStartedPulling="2026-03-13 09:51:42.385618903 +0000 UTC m=+2343.135533580" lastFinishedPulling="2026-03-13 09:51:43.017406967 +0000 UTC m=+2343.767321644" observedRunningTime="2026-03-13 09:51:43.409264724 +0000 UTC m=+2344.159179401" watchObservedRunningTime="2026-03-13 09:51:43.419704085 +0000 UTC m=+2344.169618762" Mar 13 09:51:48 crc kubenswrapper[4930]: I0313 09:51:48.971310 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:51:48 crc kubenswrapper[4930]: E0313 09:51:48.972217 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:52:00 crc kubenswrapper[4930]: I0313 09:52:00.149127 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556592-zfrrf"] Mar 13 09:52:00 crc kubenswrapper[4930]: I0313 09:52:00.151777 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556592-zfrrf" Mar 13 09:52:00 crc kubenswrapper[4930]: I0313 09:52:00.154098 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:52:00 crc kubenswrapper[4930]: I0313 09:52:00.154282 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:52:00 crc kubenswrapper[4930]: I0313 09:52:00.155642 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:52:00 crc kubenswrapper[4930]: I0313 09:52:00.167684 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556592-zfrrf"] Mar 13 09:52:00 crc kubenswrapper[4930]: I0313 09:52:00.216516 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m57l7\" (UniqueName: \"kubernetes.io/projected/44373eba-2f48-4bf4-9f3e-2b262e198007-kube-api-access-m57l7\") pod \"auto-csr-approver-29556592-zfrrf\" (UID: \"44373eba-2f48-4bf4-9f3e-2b262e198007\") " pod="openshift-infra/auto-csr-approver-29556592-zfrrf" Mar 13 09:52:00 crc kubenswrapper[4930]: I0313 09:52:00.319321 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m57l7\" (UniqueName: \"kubernetes.io/projected/44373eba-2f48-4bf4-9f3e-2b262e198007-kube-api-access-m57l7\") pod \"auto-csr-approver-29556592-zfrrf\" (UID: \"44373eba-2f48-4bf4-9f3e-2b262e198007\") " pod="openshift-infra/auto-csr-approver-29556592-zfrrf" Mar 13 09:52:00 crc kubenswrapper[4930]: I0313 09:52:00.342658 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m57l7\" (UniqueName: \"kubernetes.io/projected/44373eba-2f48-4bf4-9f3e-2b262e198007-kube-api-access-m57l7\") pod \"auto-csr-approver-29556592-zfrrf\" (UID: \"44373eba-2f48-4bf4-9f3e-2b262e198007\") " pod="openshift-infra/auto-csr-approver-29556592-zfrrf" Mar 13 09:52:00 crc kubenswrapper[4930]: I0313 09:52:00.472527 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556592-zfrrf" Mar 13 09:52:01 crc kubenswrapper[4930]: I0313 09:52:01.025794 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556592-zfrrf"] Mar 13 09:52:01 crc kubenswrapper[4930]: I0313 09:52:01.583548 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556592-zfrrf" event={"ID":"44373eba-2f48-4bf4-9f3e-2b262e198007","Type":"ContainerStarted","Data":"8ef9a07484297f835d396ed8899fa34659ef319020fe20443c07f21552134f1d"} Mar 13 09:52:02 crc kubenswrapper[4930]: I0313 09:52:02.594023 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556592-zfrrf" event={"ID":"44373eba-2f48-4bf4-9f3e-2b262e198007","Type":"ContainerStarted","Data":"de5e9e64ad77b5d593763cffb82c24899f48e39f950012ef4b01c58592590a6a"} Mar 13 09:52:02 crc kubenswrapper[4930]: I0313 09:52:02.616311 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556592-zfrrf" podStartSLOduration=1.573409761 podStartE2EDuration="2.616288761s" podCreationTimestamp="2026-03-13 09:52:00 +0000 UTC" firstStartedPulling="2026-03-13 09:52:01.032187381 +0000 UTC m=+2361.782102058" lastFinishedPulling="2026-03-13 09:52:02.075066381 +0000 UTC m=+2362.824981058" observedRunningTime="2026-03-13 09:52:02.607489272 +0000 UTC m=+2363.357403949" watchObservedRunningTime="2026-03-13 09:52:02.616288761 +0000 UTC m=+2363.366203438" Mar 13 09:52:03 crc kubenswrapper[4930]: I0313 09:52:03.608823 4930 generic.go:334] "Generic (PLEG): container finished" podID="44373eba-2f48-4bf4-9f3e-2b262e198007" containerID="de5e9e64ad77b5d593763cffb82c24899f48e39f950012ef4b01c58592590a6a" exitCode=0 Mar 13 09:52:03 crc kubenswrapper[4930]: I0313 09:52:03.608886 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556592-zfrrf" event={"ID":"44373eba-2f48-4bf4-9f3e-2b262e198007","Type":"ContainerDied","Data":"de5e9e64ad77b5d593763cffb82c24899f48e39f950012ef4b01c58592590a6a"} Mar 13 09:52:03 crc kubenswrapper[4930]: I0313 09:52:03.971173 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:52:03 crc kubenswrapper[4930]: E0313 09:52:03.971935 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:52:05 crc kubenswrapper[4930]: I0313 09:52:05.029281 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556592-zfrrf" Mar 13 09:52:05 crc kubenswrapper[4930]: I0313 09:52:05.147337 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m57l7\" (UniqueName: \"kubernetes.io/projected/44373eba-2f48-4bf4-9f3e-2b262e198007-kube-api-access-m57l7\") pod \"44373eba-2f48-4bf4-9f3e-2b262e198007\" (UID: \"44373eba-2f48-4bf4-9f3e-2b262e198007\") " Mar 13 09:52:05 crc kubenswrapper[4930]: I0313 09:52:05.152962 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44373eba-2f48-4bf4-9f3e-2b262e198007-kube-api-access-m57l7" (OuterVolumeSpecName: "kube-api-access-m57l7") pod "44373eba-2f48-4bf4-9f3e-2b262e198007" (UID: "44373eba-2f48-4bf4-9f3e-2b262e198007"). InnerVolumeSpecName "kube-api-access-m57l7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:52:05 crc kubenswrapper[4930]: I0313 09:52:05.250065 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m57l7\" (UniqueName: \"kubernetes.io/projected/44373eba-2f48-4bf4-9f3e-2b262e198007-kube-api-access-m57l7\") on node \"crc\" DevicePath \"\"" Mar 13 09:52:05 crc kubenswrapper[4930]: I0313 09:52:05.629157 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556592-zfrrf" event={"ID":"44373eba-2f48-4bf4-9f3e-2b262e198007","Type":"ContainerDied","Data":"8ef9a07484297f835d396ed8899fa34659ef319020fe20443c07f21552134f1d"} Mar 13 09:52:05 crc kubenswrapper[4930]: I0313 09:52:05.629207 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8ef9a07484297f835d396ed8899fa34659ef319020fe20443c07f21552134f1d" Mar 13 09:52:05 crc kubenswrapper[4930]: I0313 09:52:05.629266 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556592-zfrrf" Mar 13 09:52:05 crc kubenswrapper[4930]: I0313 09:52:05.676490 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556586-v4txf"] Mar 13 09:52:05 crc kubenswrapper[4930]: I0313 09:52:05.688908 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556586-v4txf"] Mar 13 09:52:05 crc kubenswrapper[4930]: I0313 09:52:05.984695 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cc4b291-7e47-46d6-b8f2-400d377bb246" path="/var/lib/kubelet/pods/3cc4b291-7e47-46d6-b8f2-400d377bb246/volumes" Mar 13 09:52:18 crc kubenswrapper[4930]: I0313 09:52:18.971709 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:52:18 crc kubenswrapper[4930]: E0313 09:52:18.972577 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:52:31 crc kubenswrapper[4930]: I0313 09:52:31.034205 4930 scope.go:117] "RemoveContainer" containerID="5a718ce2e52c8d0da7055d8d9a7108271368e58fe64d9f5c97880b28f9030acf" Mar 13 09:52:31 crc kubenswrapper[4930]: I0313 09:52:31.971085 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:52:31 crc kubenswrapper[4930]: E0313 09:52:31.971887 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:52:46 crc kubenswrapper[4930]: I0313 09:52:46.971470 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:52:46 crc kubenswrapper[4930]: E0313 09:52:46.972329 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:53:00 crc kubenswrapper[4930]: I0313 09:53:00.972778 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:53:00 crc kubenswrapper[4930]: E0313 09:53:00.979117 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:53:14 crc kubenswrapper[4930]: I0313 09:53:14.971322 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:53:14 crc kubenswrapper[4930]: E0313 09:53:14.972318 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:53:27 crc kubenswrapper[4930]: I0313 09:53:27.971893 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:53:27 crc kubenswrapper[4930]: E0313 09:53:27.972805 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:53:40 crc kubenswrapper[4930]: I0313 09:53:40.971341 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:53:40 crc kubenswrapper[4930]: E0313 09:53:40.972212 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:53:55 crc kubenswrapper[4930]: I0313 09:53:55.972510 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:53:55 crc kubenswrapper[4930]: E0313 09:53:55.973950 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.152059 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556594-58bkw"] Mar 13 09:54:00 crc kubenswrapper[4930]: E0313 09:54:00.153172 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44373eba-2f48-4bf4-9f3e-2b262e198007" containerName="oc" Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.153185 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="44373eba-2f48-4bf4-9f3e-2b262e198007" containerName="oc" Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.153388 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="44373eba-2f48-4bf4-9f3e-2b262e198007" containerName="oc" Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.154213 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556594-58bkw" Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.156714 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.157056 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.157188 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.168246 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556594-58bkw"] Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.223846 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz6sb\" (UniqueName: \"kubernetes.io/projected/fa238fab-8220-49a0-8a45-a4d962f23b27-kube-api-access-tz6sb\") pod \"auto-csr-approver-29556594-58bkw\" (UID: \"fa238fab-8220-49a0-8a45-a4d962f23b27\") " pod="openshift-infra/auto-csr-approver-29556594-58bkw" Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.326730 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz6sb\" (UniqueName: \"kubernetes.io/projected/fa238fab-8220-49a0-8a45-a4d962f23b27-kube-api-access-tz6sb\") pod \"auto-csr-approver-29556594-58bkw\" (UID: \"fa238fab-8220-49a0-8a45-a4d962f23b27\") " pod="openshift-infra/auto-csr-approver-29556594-58bkw" Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.344206 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz6sb\" (UniqueName: \"kubernetes.io/projected/fa238fab-8220-49a0-8a45-a4d962f23b27-kube-api-access-tz6sb\") pod \"auto-csr-approver-29556594-58bkw\" (UID: \"fa238fab-8220-49a0-8a45-a4d962f23b27\") " pod="openshift-infra/auto-csr-approver-29556594-58bkw" Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.474139 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556594-58bkw" Mar 13 09:54:00 crc kubenswrapper[4930]: I0313 09:54:00.977990 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556594-58bkw"] Mar 13 09:54:01 crc kubenswrapper[4930]: I0313 09:54:01.529370 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556594-58bkw" event={"ID":"fa238fab-8220-49a0-8a45-a4d962f23b27","Type":"ContainerStarted","Data":"1b6dd74ff16dbf97f51bbe3f99fafc2be4d1d13129037a1c21e40252ded5e0cd"} Mar 13 09:54:02 crc kubenswrapper[4930]: I0313 09:54:02.546523 4930 generic.go:334] "Generic (PLEG): container finished" podID="fa238fab-8220-49a0-8a45-a4d962f23b27" containerID="7cb71066ed32547c81506fbae8c734da982b22bffa11a7fbcd82387c8bb4bcf5" exitCode=0 Mar 13 09:54:02 crc kubenswrapper[4930]: I0313 09:54:02.546576 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556594-58bkw" event={"ID":"fa238fab-8220-49a0-8a45-a4d962f23b27","Type":"ContainerDied","Data":"7cb71066ed32547c81506fbae8c734da982b22bffa11a7fbcd82387c8bb4bcf5"} Mar 13 09:54:03 crc kubenswrapper[4930]: I0313 09:54:03.921247 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556594-58bkw" Mar 13 09:54:04 crc kubenswrapper[4930]: I0313 09:54:04.019066 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz6sb\" (UniqueName: \"kubernetes.io/projected/fa238fab-8220-49a0-8a45-a4d962f23b27-kube-api-access-tz6sb\") pod \"fa238fab-8220-49a0-8a45-a4d962f23b27\" (UID: \"fa238fab-8220-49a0-8a45-a4d962f23b27\") " Mar 13 09:54:04 crc kubenswrapper[4930]: I0313 09:54:04.024973 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa238fab-8220-49a0-8a45-a4d962f23b27-kube-api-access-tz6sb" (OuterVolumeSpecName: "kube-api-access-tz6sb") pod "fa238fab-8220-49a0-8a45-a4d962f23b27" (UID: "fa238fab-8220-49a0-8a45-a4d962f23b27"). InnerVolumeSpecName "kube-api-access-tz6sb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:54:04 crc kubenswrapper[4930]: I0313 09:54:04.123571 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz6sb\" (UniqueName: \"kubernetes.io/projected/fa238fab-8220-49a0-8a45-a4d962f23b27-kube-api-access-tz6sb\") on node \"crc\" DevicePath \"\"" Mar 13 09:54:04 crc kubenswrapper[4930]: I0313 09:54:04.581404 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556594-58bkw" event={"ID":"fa238fab-8220-49a0-8a45-a4d962f23b27","Type":"ContainerDied","Data":"1b6dd74ff16dbf97f51bbe3f99fafc2be4d1d13129037a1c21e40252ded5e0cd"} Mar 13 09:54:04 crc kubenswrapper[4930]: I0313 09:54:04.581464 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b6dd74ff16dbf97f51bbe3f99fafc2be4d1d13129037a1c21e40252ded5e0cd" Mar 13 09:54:04 crc kubenswrapper[4930]: I0313 09:54:04.581470 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556594-58bkw" Mar 13 09:54:05 crc kubenswrapper[4930]: I0313 09:54:05.000039 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556588-l9kj2"] Mar 13 09:54:05 crc kubenswrapper[4930]: I0313 09:54:05.011117 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556588-l9kj2"] Mar 13 09:54:05 crc kubenswrapper[4930]: I0313 09:54:05.984905 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08e202d3-7e4e-40e8-aa35-c3579e1b8f2d" path="/var/lib/kubelet/pods/08e202d3-7e4e-40e8-aa35-c3579e1b8f2d/volumes" Mar 13 09:54:06 crc kubenswrapper[4930]: I0313 09:54:06.971325 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:54:06 crc kubenswrapper[4930]: E0313 09:54:06.971672 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:54:21 crc kubenswrapper[4930]: I0313 09:54:21.970934 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:54:21 crc kubenswrapper[4930]: E0313 09:54:21.971742 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:54:31 crc kubenswrapper[4930]: I0313 09:54:31.138082 4930 scope.go:117] "RemoveContainer" containerID="b570600fe74e7d4c56540dda5e350fd1e7efd08ca6e5e1b86c8e8315c9f83c9e" Mar 13 09:54:34 crc kubenswrapper[4930]: I0313 09:54:34.971930 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:54:34 crc kubenswrapper[4930]: E0313 09:54:34.973208 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:54:46 crc kubenswrapper[4930]: I0313 09:54:46.971639 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:54:46 crc kubenswrapper[4930]: E0313 09:54:46.973684 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:54:58 crc kubenswrapper[4930]: I0313 09:54:58.971577 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:54:58 crc kubenswrapper[4930]: E0313 09:54:58.972310 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:55:09 crc kubenswrapper[4930]: I0313 09:55:09.471201 4930 generic.go:334] "Generic (PLEG): container finished" podID="83600dff-ecc4-44e1-ba5c-c0cb1c0363f1" containerID="df63cf641d5abb8fc04f8ac933f6d5119fbbaeb369daa43ee46a4ffee07137c5" exitCode=0 Mar 13 09:55:09 crc kubenswrapper[4930]: I0313 09:55:09.471284 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" event={"ID":"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1","Type":"ContainerDied","Data":"df63cf641d5abb8fc04f8ac933f6d5119fbbaeb369daa43ee46a4ffee07137c5"} Mar 13 09:55:10 crc kubenswrapper[4930]: I0313 09:55:10.932773 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:55:10 crc kubenswrapper[4930]: I0313 09:55:10.973185 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:55:10 crc kubenswrapper[4930]: E0313 09:55:10.973753 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.004838 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-inventory\") pod \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.004963 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-libvirt-combined-ca-bundle\") pod \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.005016 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-libvirt-secret-0\") pod \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.005144 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d448r\" (UniqueName: \"kubernetes.io/projected/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-kube-api-access-d448r\") pod \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.005185 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-ssh-key-openstack-edpm-ipam\") pod \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\" (UID: \"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1\") " Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.011677 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-kube-api-access-d448r" (OuterVolumeSpecName: "kube-api-access-d448r") pod "83600dff-ecc4-44e1-ba5c-c0cb1c0363f1" (UID: "83600dff-ecc4-44e1-ba5c-c0cb1c0363f1"). InnerVolumeSpecName "kube-api-access-d448r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.011700 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "83600dff-ecc4-44e1-ba5c-c0cb1c0363f1" (UID: "83600dff-ecc4-44e1-ba5c-c0cb1c0363f1"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.039064 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "83600dff-ecc4-44e1-ba5c-c0cb1c0363f1" (UID: "83600dff-ecc4-44e1-ba5c-c0cb1c0363f1"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.040914 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "83600dff-ecc4-44e1-ba5c-c0cb1c0363f1" (UID: "83600dff-ecc4-44e1-ba5c-c0cb1c0363f1"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.041491 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-inventory" (OuterVolumeSpecName: "inventory") pod "83600dff-ecc4-44e1-ba5c-c0cb1c0363f1" (UID: "83600dff-ecc4-44e1-ba5c-c0cb1c0363f1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.109976 4930 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.110614 4930 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.110633 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d448r\" (UniqueName: \"kubernetes.io/projected/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-kube-api-access-d448r\") on node \"crc\" DevicePath \"\"" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.110651 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.110667 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/83600dff-ecc4-44e1-ba5c-c0cb1c0363f1-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.492190 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" event={"ID":"83600dff-ecc4-44e1-ba5c-c0cb1c0363f1","Type":"ContainerDied","Data":"5fd0541ec729d7afea54f67407ce0cb0a0f4692803f9e6c7fea93eb00e13ad78"} Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.492230 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fd0541ec729d7afea54f67407ce0cb0a0f4692803f9e6c7fea93eb00e13ad78" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.492261 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-6sghf" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.607428 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f"] Mar 13 09:55:11 crc kubenswrapper[4930]: E0313 09:55:11.608006 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83600dff-ecc4-44e1-ba5c-c0cb1c0363f1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.608032 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="83600dff-ecc4-44e1-ba5c-c0cb1c0363f1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 13 09:55:11 crc kubenswrapper[4930]: E0313 09:55:11.608078 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa238fab-8220-49a0-8a45-a4d962f23b27" containerName="oc" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.608089 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa238fab-8220-49a0-8a45-a4d962f23b27" containerName="oc" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.608352 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="83600dff-ecc4-44e1-ba5c-c0cb1c0363f1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.608393 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa238fab-8220-49a0-8a45-a4d962f23b27" containerName="oc" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.609341 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.612449 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.612528 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.612451 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.612617 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.612961 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.613073 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.617064 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.625742 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f"] Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.725536 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9lsv\" (UniqueName: \"kubernetes.io/projected/1da9a4d4-a675-4822-baec-dad6fe247ee3-kube-api-access-w9lsv\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.725627 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.725651 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.725709 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.725782 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.725997 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.726068 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.726340 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.726591 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.726661 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.726827 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.829399 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.829511 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.829588 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.829611 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.829646 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.829734 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9lsv\" (UniqueName: \"kubernetes.io/projected/1da9a4d4-a675-4822-baec-dad6fe247ee3-kube-api-access-w9lsv\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.829812 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.829836 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.829874 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.829897 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.829951 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.830835 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.833609 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.833788 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.835444 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.835655 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.836188 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.836282 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.836926 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.839138 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.843748 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.847643 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9lsv\" (UniqueName: \"kubernetes.io/projected/1da9a4d4-a675-4822-baec-dad6fe247ee3-kube-api-access-w9lsv\") pod \"nova-edpm-deployment-openstack-edpm-ipam-l757f\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:11 crc kubenswrapper[4930]: I0313 09:55:11.930144 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:55:12 crc kubenswrapper[4930]: I0313 09:55:12.525580 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f"] Mar 13 09:55:13 crc kubenswrapper[4930]: I0313 09:55:13.514179 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" event={"ID":"1da9a4d4-a675-4822-baec-dad6fe247ee3","Type":"ContainerStarted","Data":"8807714821f3b84bd7fc1575288a5ad1a13ee2b1f1957156ece18bc9d8ab4575"} Mar 13 09:55:14 crc kubenswrapper[4930]: I0313 09:55:14.525717 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" event={"ID":"1da9a4d4-a675-4822-baec-dad6fe247ee3","Type":"ContainerStarted","Data":"9dd97313b2193279c5d5f4cf4e02d3b2509e3908acf72e6c67aa5b1e84526390"} Mar 13 09:55:14 crc kubenswrapper[4930]: I0313 09:55:14.547637 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" podStartSLOduration=2.796886355 podStartE2EDuration="3.547620055s" podCreationTimestamp="2026-03-13 09:55:11 +0000 UTC" firstStartedPulling="2026-03-13 09:55:12.525396584 +0000 UTC m=+2553.275311261" lastFinishedPulling="2026-03-13 09:55:13.276130284 +0000 UTC m=+2554.026044961" observedRunningTime="2026-03-13 09:55:14.545899021 +0000 UTC m=+2555.295813698" watchObservedRunningTime="2026-03-13 09:55:14.547620055 +0000 UTC m=+2555.297534732" Mar 13 09:55:21 crc kubenswrapper[4930]: I0313 09:55:21.971877 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:55:22 crc kubenswrapper[4930]: I0313 09:55:22.614615 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"5fdfdf4f6031ee1e4c460c1631bfff6cd403d76eae7f5fa87795d52154b3db24"} Mar 13 09:56:00 crc kubenswrapper[4930]: I0313 09:56:00.146792 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556596-tz96h"] Mar 13 09:56:00 crc kubenswrapper[4930]: I0313 09:56:00.148953 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556596-tz96h" Mar 13 09:56:00 crc kubenswrapper[4930]: I0313 09:56:00.152111 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:56:00 crc kubenswrapper[4930]: I0313 09:56:00.152171 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:56:00 crc kubenswrapper[4930]: I0313 09:56:00.152532 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:56:00 crc kubenswrapper[4930]: I0313 09:56:00.169042 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556596-tz96h"] Mar 13 09:56:00 crc kubenswrapper[4930]: I0313 09:56:00.230271 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j77fr\" (UniqueName: \"kubernetes.io/projected/597b92ee-ba9b-4e35-b135-4eb65be5a6f3-kube-api-access-j77fr\") pod \"auto-csr-approver-29556596-tz96h\" (UID: \"597b92ee-ba9b-4e35-b135-4eb65be5a6f3\") " pod="openshift-infra/auto-csr-approver-29556596-tz96h" Mar 13 09:56:00 crc kubenswrapper[4930]: I0313 09:56:00.332446 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j77fr\" (UniqueName: \"kubernetes.io/projected/597b92ee-ba9b-4e35-b135-4eb65be5a6f3-kube-api-access-j77fr\") pod \"auto-csr-approver-29556596-tz96h\" (UID: \"597b92ee-ba9b-4e35-b135-4eb65be5a6f3\") " pod="openshift-infra/auto-csr-approver-29556596-tz96h" Mar 13 09:56:00 crc kubenswrapper[4930]: I0313 09:56:00.353265 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j77fr\" (UniqueName: \"kubernetes.io/projected/597b92ee-ba9b-4e35-b135-4eb65be5a6f3-kube-api-access-j77fr\") pod \"auto-csr-approver-29556596-tz96h\" (UID: \"597b92ee-ba9b-4e35-b135-4eb65be5a6f3\") " pod="openshift-infra/auto-csr-approver-29556596-tz96h" Mar 13 09:56:00 crc kubenswrapper[4930]: I0313 09:56:00.469778 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556596-tz96h" Mar 13 09:56:00 crc kubenswrapper[4930]: I0313 09:56:00.936955 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556596-tz96h"] Mar 13 09:56:00 crc kubenswrapper[4930]: I0313 09:56:00.944823 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 09:56:01 crc kubenswrapper[4930]: I0313 09:56:01.039328 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556596-tz96h" event={"ID":"597b92ee-ba9b-4e35-b135-4eb65be5a6f3","Type":"ContainerStarted","Data":"c3b29737013297f0f46ef9226b05c33d1d7480a3af17ea9920a1227f7634d767"} Mar 13 09:56:03 crc kubenswrapper[4930]: I0313 09:56:03.066846 4930 generic.go:334] "Generic (PLEG): container finished" podID="597b92ee-ba9b-4e35-b135-4eb65be5a6f3" containerID="0a90d80cd299121a5060eb9418b2347309a4e4fdcae71860d7a2668580158a5d" exitCode=0 Mar 13 09:56:03 crc kubenswrapper[4930]: I0313 09:56:03.066920 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556596-tz96h" event={"ID":"597b92ee-ba9b-4e35-b135-4eb65be5a6f3","Type":"ContainerDied","Data":"0a90d80cd299121a5060eb9418b2347309a4e4fdcae71860d7a2668580158a5d"} Mar 13 09:56:04 crc kubenswrapper[4930]: I0313 09:56:04.490941 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556596-tz96h" Mar 13 09:56:04 crc kubenswrapper[4930]: I0313 09:56:04.540824 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j77fr\" (UniqueName: \"kubernetes.io/projected/597b92ee-ba9b-4e35-b135-4eb65be5a6f3-kube-api-access-j77fr\") pod \"597b92ee-ba9b-4e35-b135-4eb65be5a6f3\" (UID: \"597b92ee-ba9b-4e35-b135-4eb65be5a6f3\") " Mar 13 09:56:04 crc kubenswrapper[4930]: I0313 09:56:04.550033 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/597b92ee-ba9b-4e35-b135-4eb65be5a6f3-kube-api-access-j77fr" (OuterVolumeSpecName: "kube-api-access-j77fr") pod "597b92ee-ba9b-4e35-b135-4eb65be5a6f3" (UID: "597b92ee-ba9b-4e35-b135-4eb65be5a6f3"). InnerVolumeSpecName "kube-api-access-j77fr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:56:04 crc kubenswrapper[4930]: I0313 09:56:04.644565 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j77fr\" (UniqueName: \"kubernetes.io/projected/597b92ee-ba9b-4e35-b135-4eb65be5a6f3-kube-api-access-j77fr\") on node \"crc\" DevicePath \"\"" Mar 13 09:56:05 crc kubenswrapper[4930]: I0313 09:56:05.087190 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556596-tz96h" event={"ID":"597b92ee-ba9b-4e35-b135-4eb65be5a6f3","Type":"ContainerDied","Data":"c3b29737013297f0f46ef9226b05c33d1d7480a3af17ea9920a1227f7634d767"} Mar 13 09:56:05 crc kubenswrapper[4930]: I0313 09:56:05.087545 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c3b29737013297f0f46ef9226b05c33d1d7480a3af17ea9920a1227f7634d767" Mar 13 09:56:05 crc kubenswrapper[4930]: I0313 09:56:05.087233 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556596-tz96h" Mar 13 09:56:05 crc kubenswrapper[4930]: I0313 09:56:05.558420 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556590-hnzrg"] Mar 13 09:56:05 crc kubenswrapper[4930]: I0313 09:56:05.571224 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556590-hnzrg"] Mar 13 09:56:05 crc kubenswrapper[4930]: I0313 09:56:05.987077 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="abd4c5da-c943-4cd3-9138-7b5c0cc39aef" path="/var/lib/kubelet/pods/abd4c5da-c943-4cd3-9138-7b5c0cc39aef/volumes" Mar 13 09:56:31 crc kubenswrapper[4930]: I0313 09:56:31.262703 4930 scope.go:117] "RemoveContainer" containerID="6e93b7d86ada97ba1bd5dfa206bcdca731cac44253657583d74c90eeaf1244d9" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.136490 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-swwh2"] Mar 13 09:57:12 crc kubenswrapper[4930]: E0313 09:57:12.137792 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="597b92ee-ba9b-4e35-b135-4eb65be5a6f3" containerName="oc" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.137808 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="597b92ee-ba9b-4e35-b135-4eb65be5a6f3" containerName="oc" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.138077 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="597b92ee-ba9b-4e35-b135-4eb65be5a6f3" containerName="oc" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.141248 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.148938 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swwh2"] Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.210103 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ec8045-ae5d-47df-a104-5a1a22e669be-catalog-content\") pod \"community-operators-swwh2\" (UID: \"47ec8045-ae5d-47df-a104-5a1a22e669be\") " pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.210147 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ec8045-ae5d-47df-a104-5a1a22e669be-utilities\") pod \"community-operators-swwh2\" (UID: \"47ec8045-ae5d-47df-a104-5a1a22e669be\") " pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.210589 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnrsz\" (UniqueName: \"kubernetes.io/projected/47ec8045-ae5d-47df-a104-5a1a22e669be-kube-api-access-xnrsz\") pod \"community-operators-swwh2\" (UID: \"47ec8045-ae5d-47df-a104-5a1a22e669be\") " pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.313347 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnrsz\" (UniqueName: \"kubernetes.io/projected/47ec8045-ae5d-47df-a104-5a1a22e669be-kube-api-access-xnrsz\") pod \"community-operators-swwh2\" (UID: \"47ec8045-ae5d-47df-a104-5a1a22e669be\") " pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.313602 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ec8045-ae5d-47df-a104-5a1a22e669be-catalog-content\") pod \"community-operators-swwh2\" (UID: \"47ec8045-ae5d-47df-a104-5a1a22e669be\") " pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.313638 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ec8045-ae5d-47df-a104-5a1a22e669be-utilities\") pod \"community-operators-swwh2\" (UID: \"47ec8045-ae5d-47df-a104-5a1a22e669be\") " pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.314062 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ec8045-ae5d-47df-a104-5a1a22e669be-catalog-content\") pod \"community-operators-swwh2\" (UID: \"47ec8045-ae5d-47df-a104-5a1a22e669be\") " pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.314171 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ec8045-ae5d-47df-a104-5a1a22e669be-utilities\") pod \"community-operators-swwh2\" (UID: \"47ec8045-ae5d-47df-a104-5a1a22e669be\") " pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.332810 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnrsz\" (UniqueName: \"kubernetes.io/projected/47ec8045-ae5d-47df-a104-5a1a22e669be-kube-api-access-xnrsz\") pod \"community-operators-swwh2\" (UID: \"47ec8045-ae5d-47df-a104-5a1a22e669be\") " pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:12 crc kubenswrapper[4930]: I0313 09:57:12.463596 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:13 crc kubenswrapper[4930]: I0313 09:57:13.134413 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swwh2"] Mar 13 09:57:14 crc kubenswrapper[4930]: I0313 09:57:14.005057 4930 generic.go:334] "Generic (PLEG): container finished" podID="47ec8045-ae5d-47df-a104-5a1a22e669be" containerID="c592b1579b7f0a15ce1614a987b0bd0bacdb2d80dbdf2b3afbda4cac9f5e673c" exitCode=0 Mar 13 09:57:14 crc kubenswrapper[4930]: I0313 09:57:14.005146 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwh2" event={"ID":"47ec8045-ae5d-47df-a104-5a1a22e669be","Type":"ContainerDied","Data":"c592b1579b7f0a15ce1614a987b0bd0bacdb2d80dbdf2b3afbda4cac9f5e673c"} Mar 13 09:57:14 crc kubenswrapper[4930]: I0313 09:57:14.005858 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwh2" event={"ID":"47ec8045-ae5d-47df-a104-5a1a22e669be","Type":"ContainerStarted","Data":"db54138324c853ba5c2e8f58563d89db6c587c5e773baa44eed0a855aaa8f67d"} Mar 13 09:57:15 crc kubenswrapper[4930]: I0313 09:57:15.023487 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwh2" event={"ID":"47ec8045-ae5d-47df-a104-5a1a22e669be","Type":"ContainerStarted","Data":"27290a42e5c9df3d4d1fb98c94fd6295e18927bddbb6a79fc6ba73a3dd2541fe"} Mar 13 09:57:17 crc kubenswrapper[4930]: I0313 09:57:17.045945 4930 generic.go:334] "Generic (PLEG): container finished" podID="47ec8045-ae5d-47df-a104-5a1a22e669be" containerID="27290a42e5c9df3d4d1fb98c94fd6295e18927bddbb6a79fc6ba73a3dd2541fe" exitCode=0 Mar 13 09:57:17 crc kubenswrapper[4930]: I0313 09:57:17.046031 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwh2" event={"ID":"47ec8045-ae5d-47df-a104-5a1a22e669be","Type":"ContainerDied","Data":"27290a42e5c9df3d4d1fb98c94fd6295e18927bddbb6a79fc6ba73a3dd2541fe"} Mar 13 09:57:18 crc kubenswrapper[4930]: I0313 09:57:18.094563 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwh2" event={"ID":"47ec8045-ae5d-47df-a104-5a1a22e669be","Type":"ContainerStarted","Data":"8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2"} Mar 13 09:57:18 crc kubenswrapper[4930]: I0313 09:57:18.121545 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-swwh2" podStartSLOduration=2.638274738 podStartE2EDuration="6.121527433s" podCreationTimestamp="2026-03-13 09:57:12 +0000 UTC" firstStartedPulling="2026-03-13 09:57:14.00960982 +0000 UTC m=+2674.759524497" lastFinishedPulling="2026-03-13 09:57:17.492862515 +0000 UTC m=+2678.242777192" observedRunningTime="2026-03-13 09:57:18.113395639 +0000 UTC m=+2678.863310316" watchObservedRunningTime="2026-03-13 09:57:18.121527433 +0000 UTC m=+2678.871442110" Mar 13 09:57:22 crc kubenswrapper[4930]: I0313 09:57:22.464327 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:22 crc kubenswrapper[4930]: I0313 09:57:22.465069 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:22 crc kubenswrapper[4930]: I0313 09:57:22.510499 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:23 crc kubenswrapper[4930]: I0313 09:57:23.207868 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:23 crc kubenswrapper[4930]: I0313 09:57:23.262879 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swwh2"] Mar 13 09:57:25 crc kubenswrapper[4930]: I0313 09:57:25.171867 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-swwh2" podUID="47ec8045-ae5d-47df-a104-5a1a22e669be" containerName="registry-server" containerID="cri-o://8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2" gracePeriod=2 Mar 13 09:57:25 crc kubenswrapper[4930]: I0313 09:57:25.752844 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:25 crc kubenswrapper[4930]: I0313 09:57:25.901299 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ec8045-ae5d-47df-a104-5a1a22e669be-utilities\") pod \"47ec8045-ae5d-47df-a104-5a1a22e669be\" (UID: \"47ec8045-ae5d-47df-a104-5a1a22e669be\") " Mar 13 09:57:25 crc kubenswrapper[4930]: I0313 09:57:25.901497 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ec8045-ae5d-47df-a104-5a1a22e669be-catalog-content\") pod \"47ec8045-ae5d-47df-a104-5a1a22e669be\" (UID: \"47ec8045-ae5d-47df-a104-5a1a22e669be\") " Mar 13 09:57:25 crc kubenswrapper[4930]: I0313 09:57:25.901557 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xnrsz\" (UniqueName: \"kubernetes.io/projected/47ec8045-ae5d-47df-a104-5a1a22e669be-kube-api-access-xnrsz\") pod \"47ec8045-ae5d-47df-a104-5a1a22e669be\" (UID: \"47ec8045-ae5d-47df-a104-5a1a22e669be\") " Mar 13 09:57:25 crc kubenswrapper[4930]: I0313 09:57:25.902152 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47ec8045-ae5d-47df-a104-5a1a22e669be-utilities" (OuterVolumeSpecName: "utilities") pod "47ec8045-ae5d-47df-a104-5a1a22e669be" (UID: "47ec8045-ae5d-47df-a104-5a1a22e669be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:57:25 crc kubenswrapper[4930]: I0313 09:57:25.912691 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47ec8045-ae5d-47df-a104-5a1a22e669be-kube-api-access-xnrsz" (OuterVolumeSpecName: "kube-api-access-xnrsz") pod "47ec8045-ae5d-47df-a104-5a1a22e669be" (UID: "47ec8045-ae5d-47df-a104-5a1a22e669be"). InnerVolumeSpecName "kube-api-access-xnrsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:57:25 crc kubenswrapper[4930]: I0313 09:57:25.970589 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47ec8045-ae5d-47df-a104-5a1a22e669be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47ec8045-ae5d-47df-a104-5a1a22e669be" (UID: "47ec8045-ae5d-47df-a104-5a1a22e669be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.005089 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ec8045-ae5d-47df-a104-5a1a22e669be-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.005139 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ec8045-ae5d-47df-a104-5a1a22e669be-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.005156 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xnrsz\" (UniqueName: \"kubernetes.io/projected/47ec8045-ae5d-47df-a104-5a1a22e669be-kube-api-access-xnrsz\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.183559 4930 generic.go:334] "Generic (PLEG): container finished" podID="47ec8045-ae5d-47df-a104-5a1a22e669be" containerID="8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2" exitCode=0 Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.183746 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwh2" event={"ID":"47ec8045-ae5d-47df-a104-5a1a22e669be","Type":"ContainerDied","Data":"8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2"} Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.183944 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swwh2" event={"ID":"47ec8045-ae5d-47df-a104-5a1a22e669be","Type":"ContainerDied","Data":"db54138324c853ba5c2e8f58563d89db6c587c5e773baa44eed0a855aaa8f67d"} Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.183963 4930 scope.go:117] "RemoveContainer" containerID="8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.183817 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swwh2" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.211887 4930 scope.go:117] "RemoveContainer" containerID="27290a42e5c9df3d4d1fb98c94fd6295e18927bddbb6a79fc6ba73a3dd2541fe" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.214614 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swwh2"] Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.227173 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-swwh2"] Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.235236 4930 scope.go:117] "RemoveContainer" containerID="c592b1579b7f0a15ce1614a987b0bd0bacdb2d80dbdf2b3afbda4cac9f5e673c" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.290343 4930 scope.go:117] "RemoveContainer" containerID="8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2" Mar 13 09:57:26 crc kubenswrapper[4930]: E0313 09:57:26.290801 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2\": container with ID starting with 8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2 not found: ID does not exist" containerID="8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.290848 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2"} err="failed to get container status \"8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2\": rpc error: code = NotFound desc = could not find container \"8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2\": container with ID starting with 8cdaade0287c6092c3d5595e2b3f5ca65360e69a5ddd06b5b7dc773dbb221bd2 not found: ID does not exist" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.290878 4930 scope.go:117] "RemoveContainer" containerID="27290a42e5c9df3d4d1fb98c94fd6295e18927bddbb6a79fc6ba73a3dd2541fe" Mar 13 09:57:26 crc kubenswrapper[4930]: E0313 09:57:26.291173 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27290a42e5c9df3d4d1fb98c94fd6295e18927bddbb6a79fc6ba73a3dd2541fe\": container with ID starting with 27290a42e5c9df3d4d1fb98c94fd6295e18927bddbb6a79fc6ba73a3dd2541fe not found: ID does not exist" containerID="27290a42e5c9df3d4d1fb98c94fd6295e18927bddbb6a79fc6ba73a3dd2541fe" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.291224 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27290a42e5c9df3d4d1fb98c94fd6295e18927bddbb6a79fc6ba73a3dd2541fe"} err="failed to get container status \"27290a42e5c9df3d4d1fb98c94fd6295e18927bddbb6a79fc6ba73a3dd2541fe\": rpc error: code = NotFound desc = could not find container \"27290a42e5c9df3d4d1fb98c94fd6295e18927bddbb6a79fc6ba73a3dd2541fe\": container with ID starting with 27290a42e5c9df3d4d1fb98c94fd6295e18927bddbb6a79fc6ba73a3dd2541fe not found: ID does not exist" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.291257 4930 scope.go:117] "RemoveContainer" containerID="c592b1579b7f0a15ce1614a987b0bd0bacdb2d80dbdf2b3afbda4cac9f5e673c" Mar 13 09:57:26 crc kubenswrapper[4930]: E0313 09:57:26.291763 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c592b1579b7f0a15ce1614a987b0bd0bacdb2d80dbdf2b3afbda4cac9f5e673c\": container with ID starting with c592b1579b7f0a15ce1614a987b0bd0bacdb2d80dbdf2b3afbda4cac9f5e673c not found: ID does not exist" containerID="c592b1579b7f0a15ce1614a987b0bd0bacdb2d80dbdf2b3afbda4cac9f5e673c" Mar 13 09:57:26 crc kubenswrapper[4930]: I0313 09:57:26.291788 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c592b1579b7f0a15ce1614a987b0bd0bacdb2d80dbdf2b3afbda4cac9f5e673c"} err="failed to get container status \"c592b1579b7f0a15ce1614a987b0bd0bacdb2d80dbdf2b3afbda4cac9f5e673c\": rpc error: code = NotFound desc = could not find container \"c592b1579b7f0a15ce1614a987b0bd0bacdb2d80dbdf2b3afbda4cac9f5e673c\": container with ID starting with c592b1579b7f0a15ce1614a987b0bd0bacdb2d80dbdf2b3afbda4cac9f5e673c not found: ID does not exist" Mar 13 09:57:27 crc kubenswrapper[4930]: I0313 09:57:27.984134 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47ec8045-ae5d-47df-a104-5a1a22e669be" path="/var/lib/kubelet/pods/47ec8045-ae5d-47df-a104-5a1a22e669be/volumes" Mar 13 09:57:28 crc kubenswrapper[4930]: I0313 09:57:28.218015 4930 generic.go:334] "Generic (PLEG): container finished" podID="1da9a4d4-a675-4822-baec-dad6fe247ee3" containerID="9dd97313b2193279c5d5f4cf4e02d3b2509e3908acf72e6c67aa5b1e84526390" exitCode=0 Mar 13 09:57:28 crc kubenswrapper[4930]: I0313 09:57:28.218065 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" event={"ID":"1da9a4d4-a675-4822-baec-dad6fe247ee3","Type":"ContainerDied","Data":"9dd97313b2193279c5d5f4cf4e02d3b2509e3908acf72e6c67aa5b1e84526390"} Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.800294 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.908237 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-migration-ssh-key-0\") pod \"1da9a4d4-a675-4822-baec-dad6fe247ee3\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.908422 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-migration-ssh-key-1\") pod \"1da9a4d4-a675-4822-baec-dad6fe247ee3\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.908488 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-3\") pod \"1da9a4d4-a675-4822-baec-dad6fe247ee3\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.908584 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-1\") pod \"1da9a4d4-a675-4822-baec-dad6fe247ee3\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.908644 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-2\") pod \"1da9a4d4-a675-4822-baec-dad6fe247ee3\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.908688 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-ssh-key-openstack-edpm-ipam\") pod \"1da9a4d4-a675-4822-baec-dad6fe247ee3\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.908710 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-combined-ca-bundle\") pod \"1da9a4d4-a675-4822-baec-dad6fe247ee3\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.908743 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-extra-config-0\") pod \"1da9a4d4-a675-4822-baec-dad6fe247ee3\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.908801 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-inventory\") pod \"1da9a4d4-a675-4822-baec-dad6fe247ee3\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.908848 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9lsv\" (UniqueName: \"kubernetes.io/projected/1da9a4d4-a675-4822-baec-dad6fe247ee3-kube-api-access-w9lsv\") pod \"1da9a4d4-a675-4822-baec-dad6fe247ee3\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.908871 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-0\") pod \"1da9a4d4-a675-4822-baec-dad6fe247ee3\" (UID: \"1da9a4d4-a675-4822-baec-dad6fe247ee3\") " Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.921979 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1da9a4d4-a675-4822-baec-dad6fe247ee3-kube-api-access-w9lsv" (OuterVolumeSpecName: "kube-api-access-w9lsv") pod "1da9a4d4-a675-4822-baec-dad6fe247ee3" (UID: "1da9a4d4-a675-4822-baec-dad6fe247ee3"). InnerVolumeSpecName "kube-api-access-w9lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.933721 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "1da9a4d4-a675-4822-baec-dad6fe247ee3" (UID: "1da9a4d4-a675-4822-baec-dad6fe247ee3"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.962381 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "1da9a4d4-a675-4822-baec-dad6fe247ee3" (UID: "1da9a4d4-a675-4822-baec-dad6fe247ee3"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.964121 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "1da9a4d4-a675-4822-baec-dad6fe247ee3" (UID: "1da9a4d4-a675-4822-baec-dad6fe247ee3"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.968959 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "1da9a4d4-a675-4822-baec-dad6fe247ee3" (UID: "1da9a4d4-a675-4822-baec-dad6fe247ee3"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.973284 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "1da9a4d4-a675-4822-baec-dad6fe247ee3" (UID: "1da9a4d4-a675-4822-baec-dad6fe247ee3"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.975544 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "1da9a4d4-a675-4822-baec-dad6fe247ee3" (UID: "1da9a4d4-a675-4822-baec-dad6fe247ee3"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 09:57:29 crc kubenswrapper[4930]: I0313 09:57:29.991400 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "1da9a4d4-a675-4822-baec-dad6fe247ee3" (UID: "1da9a4d4-a675-4822-baec-dad6fe247ee3"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.008035 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "1da9a4d4-a675-4822-baec-dad6fe247ee3" (UID: "1da9a4d4-a675-4822-baec-dad6fe247ee3"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.011267 4930 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.011289 4930 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.011298 4930 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.011307 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9lsv\" (UniqueName: \"kubernetes.io/projected/1da9a4d4-a675-4822-baec-dad6fe247ee3-kube-api-access-w9lsv\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.011315 4930 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.011324 4930 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.011333 4930 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.011341 4930 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.011350 4930 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.016037 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1da9a4d4-a675-4822-baec-dad6fe247ee3" (UID: "1da9a4d4-a675-4822-baec-dad6fe247ee3"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.032731 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-inventory" (OuterVolumeSpecName: "inventory") pod "1da9a4d4-a675-4822-baec-dad6fe247ee3" (UID: "1da9a4d4-a675-4822-baec-dad6fe247ee3"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.114097 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.114133 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1da9a4d4-a675-4822-baec-dad6fe247ee3-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.238103 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" event={"ID":"1da9a4d4-a675-4822-baec-dad6fe247ee3","Type":"ContainerDied","Data":"8807714821f3b84bd7fc1575288a5ad1a13ee2b1f1957156ece18bc9d8ab4575"} Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.238142 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8807714821f3b84bd7fc1575288a5ad1a13ee2b1f1957156ece18bc9d8ab4575" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.238193 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-l757f" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.339887 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28"] Mar 13 09:57:30 crc kubenswrapper[4930]: E0313 09:57:30.340492 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ec8045-ae5d-47df-a104-5a1a22e669be" containerName="registry-server" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.340513 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ec8045-ae5d-47df-a104-5a1a22e669be" containerName="registry-server" Mar 13 09:57:30 crc kubenswrapper[4930]: E0313 09:57:30.340550 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1da9a4d4-a675-4822-baec-dad6fe247ee3" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.340558 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="1da9a4d4-a675-4822-baec-dad6fe247ee3" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 13 09:57:30 crc kubenswrapper[4930]: E0313 09:57:30.340595 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ec8045-ae5d-47df-a104-5a1a22e669be" containerName="extract-utilities" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.340603 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ec8045-ae5d-47df-a104-5a1a22e669be" containerName="extract-utilities" Mar 13 09:57:30 crc kubenswrapper[4930]: E0313 09:57:30.340628 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ec8045-ae5d-47df-a104-5a1a22e669be" containerName="extract-content" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.340636 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ec8045-ae5d-47df-a104-5a1a22e669be" containerName="extract-content" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.340893 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ec8045-ae5d-47df-a104-5a1a22e669be" containerName="registry-server" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.340939 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="1da9a4d4-a675-4822-baec-dad6fe247ee3" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.341946 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.344852 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.344863 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.345661 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.347204 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.347262 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.351910 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28"] Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.523541 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.523643 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.523673 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.523749 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vhj6\" (UniqueName: \"kubernetes.io/projected/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-kube-api-access-5vhj6\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.523938 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.524073 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.524143 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.626382 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.626763 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.626822 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.626870 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vhj6\" (UniqueName: \"kubernetes.io/projected/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-kube-api-access-5vhj6\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.626930 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.626988 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.627023 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.631014 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.631319 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.631911 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.631911 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.632312 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.632418 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.645196 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vhj6\" (UniqueName: \"kubernetes.io/projected/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-kube-api-access-5vhj6\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-qrz28\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:30 crc kubenswrapper[4930]: I0313 09:57:30.692471 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:57:31 crc kubenswrapper[4930]: I0313 09:57:31.266984 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28"] Mar 13 09:57:32 crc kubenswrapper[4930]: I0313 09:57:32.261036 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" event={"ID":"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7","Type":"ContainerStarted","Data":"0662172f005d4803800925a305cfc1f8ab95f031e3de9e9a1c82558e381bd49a"} Mar 13 09:57:32 crc kubenswrapper[4930]: I0313 09:57:32.261658 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" event={"ID":"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7","Type":"ContainerStarted","Data":"d75cbfa6b352394f08752acb168c519c6037cece0157f8f4305338a7b3ab4a06"} Mar 13 09:57:32 crc kubenswrapper[4930]: I0313 09:57:32.280491 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" podStartSLOduration=1.836135584 podStartE2EDuration="2.280475659s" podCreationTimestamp="2026-03-13 09:57:30 +0000 UTC" firstStartedPulling="2026-03-13 09:57:31.269787769 +0000 UTC m=+2692.019702446" lastFinishedPulling="2026-03-13 09:57:31.714127844 +0000 UTC m=+2692.464042521" observedRunningTime="2026-03-13 09:57:32.27973449 +0000 UTC m=+2693.029649157" watchObservedRunningTime="2026-03-13 09:57:32.280475659 +0000 UTC m=+2693.030390336" Mar 13 09:57:32 crc kubenswrapper[4930]: I0313 09:57:32.949196 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6d4hl"] Mar 13 09:57:32 crc kubenswrapper[4930]: I0313 09:57:32.952402 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:32 crc kubenswrapper[4930]: I0313 09:57:32.974063 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6d4hl"] Mar 13 09:57:33 crc kubenswrapper[4930]: I0313 09:57:33.091865 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab22427-a16a-4f18-8bf3-6e11675ec843-catalog-content\") pod \"redhat-marketplace-6d4hl\" (UID: \"6ab22427-a16a-4f18-8bf3-6e11675ec843\") " pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:33 crc kubenswrapper[4930]: I0313 09:57:33.092039 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kkg2\" (UniqueName: \"kubernetes.io/projected/6ab22427-a16a-4f18-8bf3-6e11675ec843-kube-api-access-6kkg2\") pod \"redhat-marketplace-6d4hl\" (UID: \"6ab22427-a16a-4f18-8bf3-6e11675ec843\") " pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:33 crc kubenswrapper[4930]: I0313 09:57:33.092181 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab22427-a16a-4f18-8bf3-6e11675ec843-utilities\") pod \"redhat-marketplace-6d4hl\" (UID: \"6ab22427-a16a-4f18-8bf3-6e11675ec843\") " pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:33 crc kubenswrapper[4930]: I0313 09:57:33.194294 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab22427-a16a-4f18-8bf3-6e11675ec843-catalog-content\") pod \"redhat-marketplace-6d4hl\" (UID: \"6ab22427-a16a-4f18-8bf3-6e11675ec843\") " pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:33 crc kubenswrapper[4930]: I0313 09:57:33.194418 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kkg2\" (UniqueName: \"kubernetes.io/projected/6ab22427-a16a-4f18-8bf3-6e11675ec843-kube-api-access-6kkg2\") pod \"redhat-marketplace-6d4hl\" (UID: \"6ab22427-a16a-4f18-8bf3-6e11675ec843\") " pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:33 crc kubenswrapper[4930]: I0313 09:57:33.194572 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab22427-a16a-4f18-8bf3-6e11675ec843-utilities\") pod \"redhat-marketplace-6d4hl\" (UID: \"6ab22427-a16a-4f18-8bf3-6e11675ec843\") " pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:33 crc kubenswrapper[4930]: I0313 09:57:33.194851 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab22427-a16a-4f18-8bf3-6e11675ec843-catalog-content\") pod \"redhat-marketplace-6d4hl\" (UID: \"6ab22427-a16a-4f18-8bf3-6e11675ec843\") " pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:33 crc kubenswrapper[4930]: I0313 09:57:33.194926 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab22427-a16a-4f18-8bf3-6e11675ec843-utilities\") pod \"redhat-marketplace-6d4hl\" (UID: \"6ab22427-a16a-4f18-8bf3-6e11675ec843\") " pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:33 crc kubenswrapper[4930]: I0313 09:57:33.215298 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kkg2\" (UniqueName: \"kubernetes.io/projected/6ab22427-a16a-4f18-8bf3-6e11675ec843-kube-api-access-6kkg2\") pod \"redhat-marketplace-6d4hl\" (UID: \"6ab22427-a16a-4f18-8bf3-6e11675ec843\") " pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:33 crc kubenswrapper[4930]: I0313 09:57:33.302212 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:33 crc kubenswrapper[4930]: I0313 09:57:33.850633 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6d4hl"] Mar 13 09:57:33 crc kubenswrapper[4930]: W0313 09:57:33.860478 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ab22427_a16a_4f18_8bf3_6e11675ec843.slice/crio-c4ce9ba4245834103700645258f9b8c67da8626d6436f9f9104348d6bbbe5a49 WatchSource:0}: Error finding container c4ce9ba4245834103700645258f9b8c67da8626d6436f9f9104348d6bbbe5a49: Status 404 returned error can't find the container with id c4ce9ba4245834103700645258f9b8c67da8626d6436f9f9104348d6bbbe5a49 Mar 13 09:57:34 crc kubenswrapper[4930]: I0313 09:57:34.281279 4930 generic.go:334] "Generic (PLEG): container finished" podID="6ab22427-a16a-4f18-8bf3-6e11675ec843" containerID="a0ee0edcdfc0ea2ffe0305f14329604e0e19524a5861a6144649540d6f131381" exitCode=0 Mar 13 09:57:34 crc kubenswrapper[4930]: I0313 09:57:34.281340 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6d4hl" event={"ID":"6ab22427-a16a-4f18-8bf3-6e11675ec843","Type":"ContainerDied","Data":"a0ee0edcdfc0ea2ffe0305f14329604e0e19524a5861a6144649540d6f131381"} Mar 13 09:57:34 crc kubenswrapper[4930]: I0313 09:57:34.281379 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6d4hl" event={"ID":"6ab22427-a16a-4f18-8bf3-6e11675ec843","Type":"ContainerStarted","Data":"c4ce9ba4245834103700645258f9b8c67da8626d6436f9f9104348d6bbbe5a49"} Mar 13 09:57:35 crc kubenswrapper[4930]: I0313 09:57:35.293416 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6d4hl" event={"ID":"6ab22427-a16a-4f18-8bf3-6e11675ec843","Type":"ContainerStarted","Data":"d284e891a1926ffbb70aced4800f15ca1dd979962c88d397898d0694869e3cf9"} Mar 13 09:57:36 crc kubenswrapper[4930]: I0313 09:57:36.304123 4930 generic.go:334] "Generic (PLEG): container finished" podID="6ab22427-a16a-4f18-8bf3-6e11675ec843" containerID="d284e891a1926ffbb70aced4800f15ca1dd979962c88d397898d0694869e3cf9" exitCode=0 Mar 13 09:57:36 crc kubenswrapper[4930]: I0313 09:57:36.304202 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6d4hl" event={"ID":"6ab22427-a16a-4f18-8bf3-6e11675ec843","Type":"ContainerDied","Data":"d284e891a1926ffbb70aced4800f15ca1dd979962c88d397898d0694869e3cf9"} Mar 13 09:57:37 crc kubenswrapper[4930]: I0313 09:57:37.316282 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6d4hl" event={"ID":"6ab22427-a16a-4f18-8bf3-6e11675ec843","Type":"ContainerStarted","Data":"4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d"} Mar 13 09:57:37 crc kubenswrapper[4930]: I0313 09:57:37.339752 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6d4hl" podStartSLOduration=2.886572404 podStartE2EDuration="5.339728952s" podCreationTimestamp="2026-03-13 09:57:32 +0000 UTC" firstStartedPulling="2026-03-13 09:57:34.28318583 +0000 UTC m=+2695.033100507" lastFinishedPulling="2026-03-13 09:57:36.736342378 +0000 UTC m=+2697.486257055" observedRunningTime="2026-03-13 09:57:37.336555983 +0000 UTC m=+2698.086470660" watchObservedRunningTime="2026-03-13 09:57:37.339728952 +0000 UTC m=+2698.089643629" Mar 13 09:57:42 crc kubenswrapper[4930]: I0313 09:57:42.308610 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:57:42 crc kubenswrapper[4930]: I0313 09:57:42.309165 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:57:43 crc kubenswrapper[4930]: I0313 09:57:43.302785 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:43 crc kubenswrapper[4930]: I0313 09:57:43.304577 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:43 crc kubenswrapper[4930]: I0313 09:57:43.357901 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:43 crc kubenswrapper[4930]: I0313 09:57:43.425692 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:43 crc kubenswrapper[4930]: I0313 09:57:43.599648 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6d4hl"] Mar 13 09:57:45 crc kubenswrapper[4930]: I0313 09:57:45.396902 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6d4hl" podUID="6ab22427-a16a-4f18-8bf3-6e11675ec843" containerName="registry-server" containerID="cri-o://4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d" gracePeriod=2 Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.003776 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.135505 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab22427-a16a-4f18-8bf3-6e11675ec843-catalog-content\") pod \"6ab22427-a16a-4f18-8bf3-6e11675ec843\" (UID: \"6ab22427-a16a-4f18-8bf3-6e11675ec843\") " Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.135769 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab22427-a16a-4f18-8bf3-6e11675ec843-utilities\") pod \"6ab22427-a16a-4f18-8bf3-6e11675ec843\" (UID: \"6ab22427-a16a-4f18-8bf3-6e11675ec843\") " Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.135806 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kkg2\" (UniqueName: \"kubernetes.io/projected/6ab22427-a16a-4f18-8bf3-6e11675ec843-kube-api-access-6kkg2\") pod \"6ab22427-a16a-4f18-8bf3-6e11675ec843\" (UID: \"6ab22427-a16a-4f18-8bf3-6e11675ec843\") " Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.136596 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ab22427-a16a-4f18-8bf3-6e11675ec843-utilities" (OuterVolumeSpecName: "utilities") pod "6ab22427-a16a-4f18-8bf3-6e11675ec843" (UID: "6ab22427-a16a-4f18-8bf3-6e11675ec843"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.136826 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6ab22427-a16a-4f18-8bf3-6e11675ec843-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.142815 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ab22427-a16a-4f18-8bf3-6e11675ec843-kube-api-access-6kkg2" (OuterVolumeSpecName: "kube-api-access-6kkg2") pod "6ab22427-a16a-4f18-8bf3-6e11675ec843" (UID: "6ab22427-a16a-4f18-8bf3-6e11675ec843"). InnerVolumeSpecName "kube-api-access-6kkg2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.161167 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6ab22427-a16a-4f18-8bf3-6e11675ec843-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6ab22427-a16a-4f18-8bf3-6e11675ec843" (UID: "6ab22427-a16a-4f18-8bf3-6e11675ec843"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.239829 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6ab22427-a16a-4f18-8bf3-6e11675ec843-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.240975 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kkg2\" (UniqueName: \"kubernetes.io/projected/6ab22427-a16a-4f18-8bf3-6e11675ec843-kube-api-access-6kkg2\") on node \"crc\" DevicePath \"\"" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.407299 4930 generic.go:334] "Generic (PLEG): container finished" podID="6ab22427-a16a-4f18-8bf3-6e11675ec843" containerID="4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d" exitCode=0 Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.408346 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6d4hl" event={"ID":"6ab22427-a16a-4f18-8bf3-6e11675ec843","Type":"ContainerDied","Data":"4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d"} Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.408474 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6d4hl" event={"ID":"6ab22427-a16a-4f18-8bf3-6e11675ec843","Type":"ContainerDied","Data":"c4ce9ba4245834103700645258f9b8c67da8626d6436f9f9104348d6bbbe5a49"} Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.408549 4930 scope.go:117] "RemoveContainer" containerID="4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.408611 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6d4hl" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.432591 4930 scope.go:117] "RemoveContainer" containerID="d284e891a1926ffbb70aced4800f15ca1dd979962c88d397898d0694869e3cf9" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.445574 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6d4hl"] Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.459096 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6d4hl"] Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.463741 4930 scope.go:117] "RemoveContainer" containerID="a0ee0edcdfc0ea2ffe0305f14329604e0e19524a5861a6144649540d6f131381" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.521790 4930 scope.go:117] "RemoveContainer" containerID="4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d" Mar 13 09:57:46 crc kubenswrapper[4930]: E0313 09:57:46.522362 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d\": container with ID starting with 4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d not found: ID does not exist" containerID="4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.522538 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d"} err="failed to get container status \"4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d\": rpc error: code = NotFound desc = could not find container \"4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d\": container with ID starting with 4c97887cd3351038c85d07785b5c9aba16aac387bc2227ee8a7d5dece6379b1d not found: ID does not exist" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.522870 4930 scope.go:117] "RemoveContainer" containerID="d284e891a1926ffbb70aced4800f15ca1dd979962c88d397898d0694869e3cf9" Mar 13 09:57:46 crc kubenswrapper[4930]: E0313 09:57:46.523348 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d284e891a1926ffbb70aced4800f15ca1dd979962c88d397898d0694869e3cf9\": container with ID starting with d284e891a1926ffbb70aced4800f15ca1dd979962c88d397898d0694869e3cf9 not found: ID does not exist" containerID="d284e891a1926ffbb70aced4800f15ca1dd979962c88d397898d0694869e3cf9" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.523383 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d284e891a1926ffbb70aced4800f15ca1dd979962c88d397898d0694869e3cf9"} err="failed to get container status \"d284e891a1926ffbb70aced4800f15ca1dd979962c88d397898d0694869e3cf9\": rpc error: code = NotFound desc = could not find container \"d284e891a1926ffbb70aced4800f15ca1dd979962c88d397898d0694869e3cf9\": container with ID starting with d284e891a1926ffbb70aced4800f15ca1dd979962c88d397898d0694869e3cf9 not found: ID does not exist" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.523408 4930 scope.go:117] "RemoveContainer" containerID="a0ee0edcdfc0ea2ffe0305f14329604e0e19524a5861a6144649540d6f131381" Mar 13 09:57:46 crc kubenswrapper[4930]: E0313 09:57:46.523683 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0ee0edcdfc0ea2ffe0305f14329604e0e19524a5861a6144649540d6f131381\": container with ID starting with a0ee0edcdfc0ea2ffe0305f14329604e0e19524a5861a6144649540d6f131381 not found: ID does not exist" containerID="a0ee0edcdfc0ea2ffe0305f14329604e0e19524a5861a6144649540d6f131381" Mar 13 09:57:46 crc kubenswrapper[4930]: I0313 09:57:46.523709 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0ee0edcdfc0ea2ffe0305f14329604e0e19524a5861a6144649540d6f131381"} err="failed to get container status \"a0ee0edcdfc0ea2ffe0305f14329604e0e19524a5861a6144649540d6f131381\": rpc error: code = NotFound desc = could not find container \"a0ee0edcdfc0ea2ffe0305f14329604e0e19524a5861a6144649540d6f131381\": container with ID starting with a0ee0edcdfc0ea2ffe0305f14329604e0e19524a5861a6144649540d6f131381 not found: ID does not exist" Mar 13 09:57:47 crc kubenswrapper[4930]: I0313 09:57:47.987751 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ab22427-a16a-4f18-8bf3-6e11675ec843" path="/var/lib/kubelet/pods/6ab22427-a16a-4f18-8bf3-6e11675ec843/volumes" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.143305 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556598-td629"] Mar 13 09:58:00 crc kubenswrapper[4930]: E0313 09:58:00.144375 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab22427-a16a-4f18-8bf3-6e11675ec843" containerName="extract-content" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.144390 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab22427-a16a-4f18-8bf3-6e11675ec843" containerName="extract-content" Mar 13 09:58:00 crc kubenswrapper[4930]: E0313 09:58:00.144421 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab22427-a16a-4f18-8bf3-6e11675ec843" containerName="registry-server" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.144427 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab22427-a16a-4f18-8bf3-6e11675ec843" containerName="registry-server" Mar 13 09:58:00 crc kubenswrapper[4930]: E0313 09:58:00.144456 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ab22427-a16a-4f18-8bf3-6e11675ec843" containerName="extract-utilities" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.144464 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ab22427-a16a-4f18-8bf3-6e11675ec843" containerName="extract-utilities" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.144737 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ab22427-a16a-4f18-8bf3-6e11675ec843" containerName="registry-server" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.145696 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556598-td629" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.149082 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.149245 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.151980 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.153933 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j6p5\" (UniqueName: \"kubernetes.io/projected/b38a35aa-d00e-4cff-960c-c933212cdeb4-kube-api-access-6j6p5\") pod \"auto-csr-approver-29556598-td629\" (UID: \"b38a35aa-d00e-4cff-960c-c933212cdeb4\") " pod="openshift-infra/auto-csr-approver-29556598-td629" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.157783 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556598-td629"] Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.256226 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j6p5\" (UniqueName: \"kubernetes.io/projected/b38a35aa-d00e-4cff-960c-c933212cdeb4-kube-api-access-6j6p5\") pod \"auto-csr-approver-29556598-td629\" (UID: \"b38a35aa-d00e-4cff-960c-c933212cdeb4\") " pod="openshift-infra/auto-csr-approver-29556598-td629" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.285844 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j6p5\" (UniqueName: \"kubernetes.io/projected/b38a35aa-d00e-4cff-960c-c933212cdeb4-kube-api-access-6j6p5\") pod \"auto-csr-approver-29556598-td629\" (UID: \"b38a35aa-d00e-4cff-960c-c933212cdeb4\") " pod="openshift-infra/auto-csr-approver-29556598-td629" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.472050 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556598-td629" Mar 13 09:58:00 crc kubenswrapper[4930]: I0313 09:58:00.995363 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556598-td629"] Mar 13 09:58:01 crc kubenswrapper[4930]: I0313 09:58:01.067945 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556598-td629" event={"ID":"b38a35aa-d00e-4cff-960c-c933212cdeb4","Type":"ContainerStarted","Data":"f7dc378dbbc6d62cdb4422009e35b0c95913d70b7f81fa3fa21a85fb2e4f8543"} Mar 13 09:58:03 crc kubenswrapper[4930]: I0313 09:58:03.105337 4930 generic.go:334] "Generic (PLEG): container finished" podID="b38a35aa-d00e-4cff-960c-c933212cdeb4" containerID="71a0d6451cc902c4c2a78587b5efdfa26f29ecd49ce737d016a1040341cae805" exitCode=0 Mar 13 09:58:03 crc kubenswrapper[4930]: I0313 09:58:03.105462 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556598-td629" event={"ID":"b38a35aa-d00e-4cff-960c-c933212cdeb4","Type":"ContainerDied","Data":"71a0d6451cc902c4c2a78587b5efdfa26f29ecd49ce737d016a1040341cae805"} Mar 13 09:58:04 crc kubenswrapper[4930]: I0313 09:58:04.577926 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556598-td629" Mar 13 09:58:04 crc kubenswrapper[4930]: I0313 09:58:04.688729 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6j6p5\" (UniqueName: \"kubernetes.io/projected/b38a35aa-d00e-4cff-960c-c933212cdeb4-kube-api-access-6j6p5\") pod \"b38a35aa-d00e-4cff-960c-c933212cdeb4\" (UID: \"b38a35aa-d00e-4cff-960c-c933212cdeb4\") " Mar 13 09:58:04 crc kubenswrapper[4930]: I0313 09:58:04.695450 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b38a35aa-d00e-4cff-960c-c933212cdeb4-kube-api-access-6j6p5" (OuterVolumeSpecName: "kube-api-access-6j6p5") pod "b38a35aa-d00e-4cff-960c-c933212cdeb4" (UID: "b38a35aa-d00e-4cff-960c-c933212cdeb4"). InnerVolumeSpecName "kube-api-access-6j6p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:58:04 crc kubenswrapper[4930]: I0313 09:58:04.792048 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6j6p5\" (UniqueName: \"kubernetes.io/projected/b38a35aa-d00e-4cff-960c-c933212cdeb4-kube-api-access-6j6p5\") on node \"crc\" DevicePath \"\"" Mar 13 09:58:05 crc kubenswrapper[4930]: I0313 09:58:05.129405 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556598-td629" event={"ID":"b38a35aa-d00e-4cff-960c-c933212cdeb4","Type":"ContainerDied","Data":"f7dc378dbbc6d62cdb4422009e35b0c95913d70b7f81fa3fa21a85fb2e4f8543"} Mar 13 09:58:05 crc kubenswrapper[4930]: I0313 09:58:05.129970 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7dc378dbbc6d62cdb4422009e35b0c95913d70b7f81fa3fa21a85fb2e4f8543" Mar 13 09:58:05 crc kubenswrapper[4930]: I0313 09:58:05.129489 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556598-td629" Mar 13 09:58:05 crc kubenswrapper[4930]: I0313 09:58:05.651887 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556592-zfrrf"] Mar 13 09:58:05 crc kubenswrapper[4930]: I0313 09:58:05.662722 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556592-zfrrf"] Mar 13 09:58:05 crc kubenswrapper[4930]: I0313 09:58:05.985850 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44373eba-2f48-4bf4-9f3e-2b262e198007" path="/var/lib/kubelet/pods/44373eba-2f48-4bf4-9f3e-2b262e198007/volumes" Mar 13 09:58:12 crc kubenswrapper[4930]: I0313 09:58:12.308205 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:58:12 crc kubenswrapper[4930]: I0313 09:58:12.308754 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.118787 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-sk95p"] Mar 13 09:58:14 crc kubenswrapper[4930]: E0313 09:58:14.119934 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b38a35aa-d00e-4cff-960c-c933212cdeb4" containerName="oc" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.119948 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b38a35aa-d00e-4cff-960c-c933212cdeb4" containerName="oc" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.120224 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b38a35aa-d00e-4cff-960c-c933212cdeb4" containerName="oc" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.122314 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.129374 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sk95p"] Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.311546 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55bpp\" (UniqueName: \"kubernetes.io/projected/0137d906-3d34-4384-995f-4fad060d80d0-kube-api-access-55bpp\") pod \"certified-operators-sk95p\" (UID: \"0137d906-3d34-4384-995f-4fad060d80d0\") " pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.311686 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0137d906-3d34-4384-995f-4fad060d80d0-catalog-content\") pod \"certified-operators-sk95p\" (UID: \"0137d906-3d34-4384-995f-4fad060d80d0\") " pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.311842 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0137d906-3d34-4384-995f-4fad060d80d0-utilities\") pod \"certified-operators-sk95p\" (UID: \"0137d906-3d34-4384-995f-4fad060d80d0\") " pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.414608 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55bpp\" (UniqueName: \"kubernetes.io/projected/0137d906-3d34-4384-995f-4fad060d80d0-kube-api-access-55bpp\") pod \"certified-operators-sk95p\" (UID: \"0137d906-3d34-4384-995f-4fad060d80d0\") " pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.415014 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0137d906-3d34-4384-995f-4fad060d80d0-catalog-content\") pod \"certified-operators-sk95p\" (UID: \"0137d906-3d34-4384-995f-4fad060d80d0\") " pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.415056 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0137d906-3d34-4384-995f-4fad060d80d0-utilities\") pod \"certified-operators-sk95p\" (UID: \"0137d906-3d34-4384-995f-4fad060d80d0\") " pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.415621 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0137d906-3d34-4384-995f-4fad060d80d0-utilities\") pod \"certified-operators-sk95p\" (UID: \"0137d906-3d34-4384-995f-4fad060d80d0\") " pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.415619 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0137d906-3d34-4384-995f-4fad060d80d0-catalog-content\") pod \"certified-operators-sk95p\" (UID: \"0137d906-3d34-4384-995f-4fad060d80d0\") " pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.439880 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55bpp\" (UniqueName: \"kubernetes.io/projected/0137d906-3d34-4384-995f-4fad060d80d0-kube-api-access-55bpp\") pod \"certified-operators-sk95p\" (UID: \"0137d906-3d34-4384-995f-4fad060d80d0\") " pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:14 crc kubenswrapper[4930]: I0313 09:58:14.485601 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:15 crc kubenswrapper[4930]: I0313 09:58:15.063354 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-sk95p"] Mar 13 09:58:15 crc kubenswrapper[4930]: I0313 09:58:15.234591 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sk95p" event={"ID":"0137d906-3d34-4384-995f-4fad060d80d0","Type":"ContainerStarted","Data":"f52ba65108d6d281b90c921a9bfa5a07e4016daf94c94848933752618bde4f2d"} Mar 13 09:58:16 crc kubenswrapper[4930]: I0313 09:58:16.245990 4930 generic.go:334] "Generic (PLEG): container finished" podID="0137d906-3d34-4384-995f-4fad060d80d0" containerID="930dfed3dafdfd4257780bb8d682c2f6aca0c9f8394d25c5237877f9bcbda4da" exitCode=0 Mar 13 09:58:16 crc kubenswrapper[4930]: I0313 09:58:16.246091 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sk95p" event={"ID":"0137d906-3d34-4384-995f-4fad060d80d0","Type":"ContainerDied","Data":"930dfed3dafdfd4257780bb8d682c2f6aca0c9f8394d25c5237877f9bcbda4da"} Mar 13 09:58:18 crc kubenswrapper[4930]: I0313 09:58:18.267463 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sk95p" event={"ID":"0137d906-3d34-4384-995f-4fad060d80d0","Type":"ContainerStarted","Data":"889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a"} Mar 13 09:58:20 crc kubenswrapper[4930]: E0313 09:58:20.097278 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0137d906_3d34_4384_995f_4fad060d80d0.slice/crio-conmon-889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a.scope\": RecentStats: unable to find data in memory cache]" Mar 13 09:58:20 crc kubenswrapper[4930]: E0313 09:58:20.097822 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0137d906_3d34_4384_995f_4fad060d80d0.slice/crio-conmon-889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a.scope\": RecentStats: unable to find data in memory cache]" Mar 13 09:58:20 crc kubenswrapper[4930]: I0313 09:58:20.290918 4930 generic.go:334] "Generic (PLEG): container finished" podID="0137d906-3d34-4384-995f-4fad060d80d0" containerID="889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a" exitCode=0 Mar 13 09:58:20 crc kubenswrapper[4930]: I0313 09:58:20.290970 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sk95p" event={"ID":"0137d906-3d34-4384-995f-4fad060d80d0","Type":"ContainerDied","Data":"889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a"} Mar 13 09:58:21 crc kubenswrapper[4930]: I0313 09:58:21.302975 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sk95p" event={"ID":"0137d906-3d34-4384-995f-4fad060d80d0","Type":"ContainerStarted","Data":"6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132"} Mar 13 09:58:21 crc kubenswrapper[4930]: I0313 09:58:21.324755 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-sk95p" podStartSLOduration=2.858927448 podStartE2EDuration="7.324732737s" podCreationTimestamp="2026-03-13 09:58:14 +0000 UTC" firstStartedPulling="2026-03-13 09:58:16.248070808 +0000 UTC m=+2736.997985485" lastFinishedPulling="2026-03-13 09:58:20.713876097 +0000 UTC m=+2741.463790774" observedRunningTime="2026-03-13 09:58:21.319533057 +0000 UTC m=+2742.069447744" watchObservedRunningTime="2026-03-13 09:58:21.324732737 +0000 UTC m=+2742.074647414" Mar 13 09:58:24 crc kubenswrapper[4930]: I0313 09:58:24.486206 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:24 crc kubenswrapper[4930]: I0313 09:58:24.486889 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:24 crc kubenswrapper[4930]: I0313 09:58:24.540995 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:25 crc kubenswrapper[4930]: I0313 09:58:25.399692 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:25 crc kubenswrapper[4930]: I0313 09:58:25.450417 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sk95p"] Mar 13 09:58:27 crc kubenswrapper[4930]: I0313 09:58:27.359768 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-sk95p" podUID="0137d906-3d34-4384-995f-4fad060d80d0" containerName="registry-server" containerID="cri-o://6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132" gracePeriod=2 Mar 13 09:58:27 crc kubenswrapper[4930]: I0313 09:58:27.984999 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.139027 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0137d906-3d34-4384-995f-4fad060d80d0-utilities\") pod \"0137d906-3d34-4384-995f-4fad060d80d0\" (UID: \"0137d906-3d34-4384-995f-4fad060d80d0\") " Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.139513 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55bpp\" (UniqueName: \"kubernetes.io/projected/0137d906-3d34-4384-995f-4fad060d80d0-kube-api-access-55bpp\") pod \"0137d906-3d34-4384-995f-4fad060d80d0\" (UID: \"0137d906-3d34-4384-995f-4fad060d80d0\") " Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.139693 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0137d906-3d34-4384-995f-4fad060d80d0-catalog-content\") pod \"0137d906-3d34-4384-995f-4fad060d80d0\" (UID: \"0137d906-3d34-4384-995f-4fad060d80d0\") " Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.142670 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0137d906-3d34-4384-995f-4fad060d80d0-utilities" (OuterVolumeSpecName: "utilities") pod "0137d906-3d34-4384-995f-4fad060d80d0" (UID: "0137d906-3d34-4384-995f-4fad060d80d0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.147751 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0137d906-3d34-4384-995f-4fad060d80d0-kube-api-access-55bpp" (OuterVolumeSpecName: "kube-api-access-55bpp") pod "0137d906-3d34-4384-995f-4fad060d80d0" (UID: "0137d906-3d34-4384-995f-4fad060d80d0"). InnerVolumeSpecName "kube-api-access-55bpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.202974 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0137d906-3d34-4384-995f-4fad060d80d0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0137d906-3d34-4384-995f-4fad060d80d0" (UID: "0137d906-3d34-4384-995f-4fad060d80d0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.243115 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0137d906-3d34-4384-995f-4fad060d80d0-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.243168 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55bpp\" (UniqueName: \"kubernetes.io/projected/0137d906-3d34-4384-995f-4fad060d80d0-kube-api-access-55bpp\") on node \"crc\" DevicePath \"\"" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.243185 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0137d906-3d34-4384-995f-4fad060d80d0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.372575 4930 generic.go:334] "Generic (PLEG): container finished" podID="0137d906-3d34-4384-995f-4fad060d80d0" containerID="6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132" exitCode=0 Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.372618 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sk95p" event={"ID":"0137d906-3d34-4384-995f-4fad060d80d0","Type":"ContainerDied","Data":"6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132"} Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.372645 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-sk95p" event={"ID":"0137d906-3d34-4384-995f-4fad060d80d0","Type":"ContainerDied","Data":"f52ba65108d6d281b90c921a9bfa5a07e4016daf94c94848933752618bde4f2d"} Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.372662 4930 scope.go:117] "RemoveContainer" containerID="6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.374667 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-sk95p" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.421014 4930 scope.go:117] "RemoveContainer" containerID="889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.425231 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-sk95p"] Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.457753 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-sk95p"] Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.470817 4930 scope.go:117] "RemoveContainer" containerID="930dfed3dafdfd4257780bb8d682c2f6aca0c9f8394d25c5237877f9bcbda4da" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.559944 4930 scope.go:117] "RemoveContainer" containerID="6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132" Mar 13 09:58:28 crc kubenswrapper[4930]: E0313 09:58:28.565478 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132\": container with ID starting with 6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132 not found: ID does not exist" containerID="6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.565522 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132"} err="failed to get container status \"6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132\": rpc error: code = NotFound desc = could not find container \"6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132\": container with ID starting with 6387aedcbaa186c2c2ff14e6cd055d98b0978c58df1961efc3e66c2a89c97132 not found: ID does not exist" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.565549 4930 scope.go:117] "RemoveContainer" containerID="889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a" Mar 13 09:58:28 crc kubenswrapper[4930]: E0313 09:58:28.566017 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a\": container with ID starting with 889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a not found: ID does not exist" containerID="889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.566046 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a"} err="failed to get container status \"889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a\": rpc error: code = NotFound desc = could not find container \"889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a\": container with ID starting with 889dd4c2b0ee42df53c30ce51d4413f3cc1cf165280e43b9b38482866886ec8a not found: ID does not exist" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.566064 4930 scope.go:117] "RemoveContainer" containerID="930dfed3dafdfd4257780bb8d682c2f6aca0c9f8394d25c5237877f9bcbda4da" Mar 13 09:58:28 crc kubenswrapper[4930]: E0313 09:58:28.566519 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"930dfed3dafdfd4257780bb8d682c2f6aca0c9f8394d25c5237877f9bcbda4da\": container with ID starting with 930dfed3dafdfd4257780bb8d682c2f6aca0c9f8394d25c5237877f9bcbda4da not found: ID does not exist" containerID="930dfed3dafdfd4257780bb8d682c2f6aca0c9f8394d25c5237877f9bcbda4da" Mar 13 09:58:28 crc kubenswrapper[4930]: I0313 09:58:28.566629 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"930dfed3dafdfd4257780bb8d682c2f6aca0c9f8394d25c5237877f9bcbda4da"} err="failed to get container status \"930dfed3dafdfd4257780bb8d682c2f6aca0c9f8394d25c5237877f9bcbda4da\": rpc error: code = NotFound desc = could not find container \"930dfed3dafdfd4257780bb8d682c2f6aca0c9f8394d25c5237877f9bcbda4da\": container with ID starting with 930dfed3dafdfd4257780bb8d682c2f6aca0c9f8394d25c5237877f9bcbda4da not found: ID does not exist" Mar 13 09:58:29 crc kubenswrapper[4930]: I0313 09:58:29.997594 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0137d906-3d34-4384-995f-4fad060d80d0" path="/var/lib/kubelet/pods/0137d906-3d34-4384-995f-4fad060d80d0/volumes" Mar 13 09:58:31 crc kubenswrapper[4930]: I0313 09:58:31.379667 4930 scope.go:117] "RemoveContainer" containerID="de5e9e64ad77b5d593763cffb82c24899f48e39f950012ef4b01c58592590a6a" Mar 13 09:58:42 crc kubenswrapper[4930]: I0313 09:58:42.308787 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 09:58:42 crc kubenswrapper[4930]: I0313 09:58:42.309478 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 09:58:42 crc kubenswrapper[4930]: I0313 09:58:42.309529 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 09:58:42 crc kubenswrapper[4930]: I0313 09:58:42.310509 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5fdfdf4f6031ee1e4c460c1631bfff6cd403d76eae7f5fa87795d52154b3db24"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 09:58:42 crc kubenswrapper[4930]: I0313 09:58:42.310572 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://5fdfdf4f6031ee1e4c460c1631bfff6cd403d76eae7f5fa87795d52154b3db24" gracePeriod=600 Mar 13 09:58:42 crc kubenswrapper[4930]: I0313 09:58:42.522577 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="5fdfdf4f6031ee1e4c460c1631bfff6cd403d76eae7f5fa87795d52154b3db24" exitCode=0 Mar 13 09:58:42 crc kubenswrapper[4930]: I0313 09:58:42.522772 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"5fdfdf4f6031ee1e4c460c1631bfff6cd403d76eae7f5fa87795d52154b3db24"} Mar 13 09:58:42 crc kubenswrapper[4930]: I0313 09:58:42.523186 4930 scope.go:117] "RemoveContainer" containerID="9226d4d69268830746202b1c5a0f69e0f1c03c19fc6a645d5df32e59ce79561f" Mar 13 09:58:43 crc kubenswrapper[4930]: I0313 09:58:43.541876 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38"} Mar 13 09:59:44 crc kubenswrapper[4930]: I0313 09:59:44.169867 4930 generic.go:334] "Generic (PLEG): container finished" podID="ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7" containerID="0662172f005d4803800925a305cfc1f8ab95f031e3de9e9a1c82558e381bd49a" exitCode=0 Mar 13 09:59:44 crc kubenswrapper[4930]: I0313 09:59:44.169966 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" event={"ID":"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7","Type":"ContainerDied","Data":"0662172f005d4803800925a305cfc1f8ab95f031e3de9e9a1c82558e381bd49a"} Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.733005 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.829905 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-0\") pod \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.830007 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-inventory\") pod \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.830059 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ssh-key-openstack-edpm-ipam\") pod \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.830082 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-2\") pod \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.830231 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-1\") pod \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.830287 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vhj6\" (UniqueName: \"kubernetes.io/projected/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-kube-api-access-5vhj6\") pod \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.830367 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-telemetry-combined-ca-bundle\") pod \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\" (UID: \"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7\") " Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.841725 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-kube-api-access-5vhj6" (OuterVolumeSpecName: "kube-api-access-5vhj6") pod "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7" (UID: "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7"). InnerVolumeSpecName "kube-api-access-5vhj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.842901 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7" (UID: "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.862493 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7" (UID: "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.865291 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7" (UID: "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.865637 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-inventory" (OuterVolumeSpecName: "inventory") pod "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7" (UID: "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.867712 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7" (UID: "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.880753 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7" (UID: "ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.933590 4930 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.933827 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.933886 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.933945 4930 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.934032 4930 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.934101 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vhj6\" (UniqueName: \"kubernetes.io/projected/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-kube-api-access-5vhj6\") on node \"crc\" DevicePath \"\"" Mar 13 09:59:45 crc kubenswrapper[4930]: I0313 09:59:45.934162 4930 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.207830 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" event={"ID":"ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7","Type":"ContainerDied","Data":"d75cbfa6b352394f08752acb168c519c6037cece0157f8f4305338a7b3ab4a06"} Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.208173 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d75cbfa6b352394f08752acb168c519c6037cece0157f8f4305338a7b3ab4a06" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.208413 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-qrz28" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.311559 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh"] Mar 13 09:59:46 crc kubenswrapper[4930]: E0313 09:59:46.312132 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.312151 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 13 09:59:46 crc kubenswrapper[4930]: E0313 09:59:46.312171 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0137d906-3d34-4384-995f-4fad060d80d0" containerName="extract-content" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.312176 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0137d906-3d34-4384-995f-4fad060d80d0" containerName="extract-content" Mar 13 09:59:46 crc kubenswrapper[4930]: E0313 09:59:46.312191 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0137d906-3d34-4384-995f-4fad060d80d0" containerName="registry-server" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.312197 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0137d906-3d34-4384-995f-4fad060d80d0" containerName="registry-server" Mar 13 09:59:46 crc kubenswrapper[4930]: E0313 09:59:46.312221 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0137d906-3d34-4384-995f-4fad060d80d0" containerName="extract-utilities" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.312227 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0137d906-3d34-4384-995f-4fad060d80d0" containerName="extract-utilities" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.312451 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce7f1ad9-05c2-458e-848c-c5bea6e0e2b7" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.312467 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0137d906-3d34-4384-995f-4fad060d80d0" containerName="registry-server" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.313353 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.317651 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.317876 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.317880 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.318094 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.318009 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.327617 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh"] Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.344627 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7lft\" (UniqueName: \"kubernetes.io/projected/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-kube-api-access-s7lft\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.344687 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ssh-key-openstack-edpm-ipam\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.344826 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.344913 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.344971 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.345053 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.345094 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.447175 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.447266 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.447309 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.447404 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.447744 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.447890 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7lft\" (UniqueName: \"kubernetes.io/projected/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-kube-api-access-s7lft\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.447919 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ssh-key-openstack-edpm-ipam\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.452698 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.452925 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.453647 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ssh-key-openstack-edpm-ipam\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.453697 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.454709 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.455824 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.463651 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7lft\" (UniqueName: \"kubernetes.io/projected/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-kube-api-access-s7lft\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:46 crc kubenswrapper[4930]: I0313 09:59:46.636905 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 09:59:47 crc kubenswrapper[4930]: I0313 09:59:47.260190 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh"] Mar 13 09:59:48 crc kubenswrapper[4930]: I0313 09:59:48.237334 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" event={"ID":"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b","Type":"ContainerStarted","Data":"8273ab3f6f4cfeea591381dd4341d0683e5c514322b4b77d973b1d3b6b0415cc"} Mar 13 09:59:51 crc kubenswrapper[4930]: I0313 09:59:51.271450 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" event={"ID":"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b","Type":"ContainerStarted","Data":"5c30a4c82b788b21624f33089c3c898a02049354fb2e300e5beb696a071c9c74"} Mar 13 09:59:51 crc kubenswrapper[4930]: I0313 09:59:51.299898 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" podStartSLOduration=2.739346513 podStartE2EDuration="5.299876879s" podCreationTimestamp="2026-03-13 09:59:46 +0000 UTC" firstStartedPulling="2026-03-13 09:59:47.278246483 +0000 UTC m=+2828.028161160" lastFinishedPulling="2026-03-13 09:59:49.838776849 +0000 UTC m=+2830.588691526" observedRunningTime="2026-03-13 09:59:51.287766966 +0000 UTC m=+2832.037681633" watchObservedRunningTime="2026-03-13 09:59:51.299876879 +0000 UTC m=+2832.049791556" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.135805 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556600-9cv2j"] Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.137883 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556600-9cv2j" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.141106 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.141566 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.147333 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556600-9cv2j"] Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.155738 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.160914 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7"] Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.162726 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.165298 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.165576 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.189024 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7"] Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.255250 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vprwq\" (UniqueName: \"kubernetes.io/projected/a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679-kube-api-access-vprwq\") pod \"auto-csr-approver-29556600-9cv2j\" (UID: \"a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679\") " pod="openshift-infra/auto-csr-approver-29556600-9cv2j" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.255608 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16a91e32-c950-47dc-a768-8ae6f8614947-secret-volume\") pod \"collect-profiles-29556600-f74f7\" (UID: \"16a91e32-c950-47dc-a768-8ae6f8614947\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.256109 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16a91e32-c950-47dc-a768-8ae6f8614947-config-volume\") pod \"collect-profiles-29556600-f74f7\" (UID: \"16a91e32-c950-47dc-a768-8ae6f8614947\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.256188 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdzdp\" (UniqueName: \"kubernetes.io/projected/16a91e32-c950-47dc-a768-8ae6f8614947-kube-api-access-xdzdp\") pod \"collect-profiles-29556600-f74f7\" (UID: \"16a91e32-c950-47dc-a768-8ae6f8614947\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.358355 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16a91e32-c950-47dc-a768-8ae6f8614947-config-volume\") pod \"collect-profiles-29556600-f74f7\" (UID: \"16a91e32-c950-47dc-a768-8ae6f8614947\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.358446 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdzdp\" (UniqueName: \"kubernetes.io/projected/16a91e32-c950-47dc-a768-8ae6f8614947-kube-api-access-xdzdp\") pod \"collect-profiles-29556600-f74f7\" (UID: \"16a91e32-c950-47dc-a768-8ae6f8614947\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.358510 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vprwq\" (UniqueName: \"kubernetes.io/projected/a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679-kube-api-access-vprwq\") pod \"auto-csr-approver-29556600-9cv2j\" (UID: \"a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679\") " pod="openshift-infra/auto-csr-approver-29556600-9cv2j" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.358671 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16a91e32-c950-47dc-a768-8ae6f8614947-secret-volume\") pod \"collect-profiles-29556600-f74f7\" (UID: \"16a91e32-c950-47dc-a768-8ae6f8614947\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.359350 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16a91e32-c950-47dc-a768-8ae6f8614947-config-volume\") pod \"collect-profiles-29556600-f74f7\" (UID: \"16a91e32-c950-47dc-a768-8ae6f8614947\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.376475 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16a91e32-c950-47dc-a768-8ae6f8614947-secret-volume\") pod \"collect-profiles-29556600-f74f7\" (UID: \"16a91e32-c950-47dc-a768-8ae6f8614947\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.379875 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vprwq\" (UniqueName: \"kubernetes.io/projected/a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679-kube-api-access-vprwq\") pod \"auto-csr-approver-29556600-9cv2j\" (UID: \"a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679\") " pod="openshift-infra/auto-csr-approver-29556600-9cv2j" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.380023 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdzdp\" (UniqueName: \"kubernetes.io/projected/16a91e32-c950-47dc-a768-8ae6f8614947-kube-api-access-xdzdp\") pod \"collect-profiles-29556600-f74f7\" (UID: \"16a91e32-c950-47dc-a768-8ae6f8614947\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.459877 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556600-9cv2j" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.489597 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.979465 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7"] Mar 13 10:00:00 crc kubenswrapper[4930]: W0313 10:00:00.984458 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda2c3e9bb_ea9c_4cf5_a3c6_37d1e85b4679.slice/crio-529de36987be303bba955fb3509f886c35f853f21e2bf21bda7b6f374c89eee3 WatchSource:0}: Error finding container 529de36987be303bba955fb3509f886c35f853f21e2bf21bda7b6f374c89eee3: Status 404 returned error can't find the container with id 529de36987be303bba955fb3509f886c35f853f21e2bf21bda7b6f374c89eee3 Mar 13 10:00:00 crc kubenswrapper[4930]: I0313 10:00:00.990947 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556600-9cv2j"] Mar 13 10:00:01 crc kubenswrapper[4930]: I0313 10:00:01.390851 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" event={"ID":"16a91e32-c950-47dc-a768-8ae6f8614947","Type":"ContainerStarted","Data":"9f69b7accb9ac4d979e03ad48c7d7b6f091384911d36c3ff0c6322eccda59419"} Mar 13 10:00:01 crc kubenswrapper[4930]: I0313 10:00:01.392120 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" event={"ID":"16a91e32-c950-47dc-a768-8ae6f8614947","Type":"ContainerStarted","Data":"3a216484843de47fdcc8d603eeac337db416d562b72eea547e5b4d63e7c68810"} Mar 13 10:00:01 crc kubenswrapper[4930]: I0313 10:00:01.392936 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556600-9cv2j" event={"ID":"a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679","Type":"ContainerStarted","Data":"529de36987be303bba955fb3509f886c35f853f21e2bf21bda7b6f374c89eee3"} Mar 13 10:00:01 crc kubenswrapper[4930]: I0313 10:00:01.415050 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" podStartSLOduration=1.4150295370000001 podStartE2EDuration="1.415029537s" podCreationTimestamp="2026-03-13 10:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:00:01.404669588 +0000 UTC m=+2842.154584275" watchObservedRunningTime="2026-03-13 10:00:01.415029537 +0000 UTC m=+2842.164944214" Mar 13 10:00:02 crc kubenswrapper[4930]: I0313 10:00:02.407732 4930 generic.go:334] "Generic (PLEG): container finished" podID="16a91e32-c950-47dc-a768-8ae6f8614947" containerID="9f69b7accb9ac4d979e03ad48c7d7b6f091384911d36c3ff0c6322eccda59419" exitCode=0 Mar 13 10:00:02 crc kubenswrapper[4930]: I0313 10:00:02.407859 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" event={"ID":"16a91e32-c950-47dc-a768-8ae6f8614947","Type":"ContainerDied","Data":"9f69b7accb9ac4d979e03ad48c7d7b6f091384911d36c3ff0c6322eccda59419"} Mar 13 10:00:03 crc kubenswrapper[4930]: I0313 10:00:03.827181 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:03 crc kubenswrapper[4930]: I0313 10:00:03.949048 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16a91e32-c950-47dc-a768-8ae6f8614947-config-volume\") pod \"16a91e32-c950-47dc-a768-8ae6f8614947\" (UID: \"16a91e32-c950-47dc-a768-8ae6f8614947\") " Mar 13 10:00:03 crc kubenswrapper[4930]: I0313 10:00:03.949234 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdzdp\" (UniqueName: \"kubernetes.io/projected/16a91e32-c950-47dc-a768-8ae6f8614947-kube-api-access-xdzdp\") pod \"16a91e32-c950-47dc-a768-8ae6f8614947\" (UID: \"16a91e32-c950-47dc-a768-8ae6f8614947\") " Mar 13 10:00:03 crc kubenswrapper[4930]: I0313 10:00:03.949361 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16a91e32-c950-47dc-a768-8ae6f8614947-secret-volume\") pod \"16a91e32-c950-47dc-a768-8ae6f8614947\" (UID: \"16a91e32-c950-47dc-a768-8ae6f8614947\") " Mar 13 10:00:03 crc kubenswrapper[4930]: I0313 10:00:03.949884 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/16a91e32-c950-47dc-a768-8ae6f8614947-config-volume" (OuterVolumeSpecName: "config-volume") pod "16a91e32-c950-47dc-a768-8ae6f8614947" (UID: "16a91e32-c950-47dc-a768-8ae6f8614947"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:00:03 crc kubenswrapper[4930]: I0313 10:00:03.950175 4930 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/16a91e32-c950-47dc-a768-8ae6f8614947-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:00:03 crc kubenswrapper[4930]: I0313 10:00:03.956267 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/16a91e32-c950-47dc-a768-8ae6f8614947-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "16a91e32-c950-47dc-a768-8ae6f8614947" (UID: "16a91e32-c950-47dc-a768-8ae6f8614947"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:00:03 crc kubenswrapper[4930]: I0313 10:00:03.956906 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16a91e32-c950-47dc-a768-8ae6f8614947-kube-api-access-xdzdp" (OuterVolumeSpecName: "kube-api-access-xdzdp") pod "16a91e32-c950-47dc-a768-8ae6f8614947" (UID: "16a91e32-c950-47dc-a768-8ae6f8614947"). InnerVolumeSpecName "kube-api-access-xdzdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:00:04 crc kubenswrapper[4930]: I0313 10:00:04.052678 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdzdp\" (UniqueName: \"kubernetes.io/projected/16a91e32-c950-47dc-a768-8ae6f8614947-kube-api-access-xdzdp\") on node \"crc\" DevicePath \"\"" Mar 13 10:00:04 crc kubenswrapper[4930]: I0313 10:00:04.052979 4930 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/16a91e32-c950-47dc-a768-8ae6f8614947-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:00:04 crc kubenswrapper[4930]: I0313 10:00:04.440710 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" event={"ID":"16a91e32-c950-47dc-a768-8ae6f8614947","Type":"ContainerDied","Data":"3a216484843de47fdcc8d603eeac337db416d562b72eea547e5b4d63e7c68810"} Mar 13 10:00:04 crc kubenswrapper[4930]: I0313 10:00:04.441005 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a216484843de47fdcc8d603eeac337db416d562b72eea547e5b4d63e7c68810" Mar 13 10:00:04 crc kubenswrapper[4930]: I0313 10:00:04.440788 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556600-f74f7" Mar 13 10:00:04 crc kubenswrapper[4930]: I0313 10:00:04.484731 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b"] Mar 13 10:00:04 crc kubenswrapper[4930]: I0313 10:00:04.497261 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556555-svg2b"] Mar 13 10:00:05 crc kubenswrapper[4930]: I0313 10:00:05.454244 4930 generic.go:334] "Generic (PLEG): container finished" podID="a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679" containerID="b1571594ae7616433e672271e95e4c41cc8610001fa6780c29bcb98cdb1fa2e6" exitCode=0 Mar 13 10:00:05 crc kubenswrapper[4930]: I0313 10:00:05.454301 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556600-9cv2j" event={"ID":"a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679","Type":"ContainerDied","Data":"b1571594ae7616433e672271e95e4c41cc8610001fa6780c29bcb98cdb1fa2e6"} Mar 13 10:00:05 crc kubenswrapper[4930]: I0313 10:00:05.993079 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58844174-f3d8-4618-a396-78d721d893b8" path="/var/lib/kubelet/pods/58844174-f3d8-4618-a396-78d721d893b8/volumes" Mar 13 10:00:06 crc kubenswrapper[4930]: I0313 10:00:06.950429 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556600-9cv2j" Mar 13 10:00:07 crc kubenswrapper[4930]: I0313 10:00:07.043903 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vprwq\" (UniqueName: \"kubernetes.io/projected/a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679-kube-api-access-vprwq\") pod \"a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679\" (UID: \"a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679\") " Mar 13 10:00:07 crc kubenswrapper[4930]: I0313 10:00:07.058132 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679-kube-api-access-vprwq" (OuterVolumeSpecName: "kube-api-access-vprwq") pod "a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679" (UID: "a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679"). InnerVolumeSpecName "kube-api-access-vprwq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:00:07 crc kubenswrapper[4930]: I0313 10:00:07.148281 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vprwq\" (UniqueName: \"kubernetes.io/projected/a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679-kube-api-access-vprwq\") on node \"crc\" DevicePath \"\"" Mar 13 10:00:07 crc kubenswrapper[4930]: I0313 10:00:07.494831 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556600-9cv2j" event={"ID":"a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679","Type":"ContainerDied","Data":"529de36987be303bba955fb3509f886c35f853f21e2bf21bda7b6f374c89eee3"} Mar 13 10:00:07 crc kubenswrapper[4930]: I0313 10:00:07.495664 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="529de36987be303bba955fb3509f886c35f853f21e2bf21bda7b6f374c89eee3" Mar 13 10:00:07 crc kubenswrapper[4930]: I0313 10:00:07.495798 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556600-9cv2j" Mar 13 10:00:08 crc kubenswrapper[4930]: I0313 10:00:08.053620 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556594-58bkw"] Mar 13 10:00:08 crc kubenswrapper[4930]: I0313 10:00:08.071253 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556594-58bkw"] Mar 13 10:00:09 crc kubenswrapper[4930]: I0313 10:00:09.989954 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa238fab-8220-49a0-8a45-a4d962f23b27" path="/var/lib/kubelet/pods/fa238fab-8220-49a0-8a45-a4d962f23b27/volumes" Mar 13 10:00:31 crc kubenswrapper[4930]: I0313 10:00:31.517882 4930 scope.go:117] "RemoveContainer" containerID="7cb71066ed32547c81506fbae8c734da982b22bffa11a7fbcd82387c8bb4bcf5" Mar 13 10:00:31 crc kubenswrapper[4930]: I0313 10:00:31.560821 4930 scope.go:117] "RemoveContainer" containerID="818e40fcb7b2a9b3918b664f7ccb63ffebdf47c83c92db171b851960d36fd850" Mar 13 10:00:42 crc kubenswrapper[4930]: I0313 10:00:42.308318 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:00:42 crc kubenswrapper[4930]: I0313 10:00:42.308879 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.153316 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29556601-msl6s"] Mar 13 10:01:00 crc kubenswrapper[4930]: E0313 10:01:00.154509 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16a91e32-c950-47dc-a768-8ae6f8614947" containerName="collect-profiles" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.154525 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="16a91e32-c950-47dc-a768-8ae6f8614947" containerName="collect-profiles" Mar 13 10:01:00 crc kubenswrapper[4930]: E0313 10:01:00.154563 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679" containerName="oc" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.154572 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679" containerName="oc" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.154824 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="16a91e32-c950-47dc-a768-8ae6f8614947" containerName="collect-profiles" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.154860 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679" containerName="oc" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.155890 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.164744 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29556601-msl6s"] Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.303041 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-combined-ca-bundle\") pod \"keystone-cron-29556601-msl6s\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.303142 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-fernet-keys\") pod \"keystone-cron-29556601-msl6s\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.303611 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-config-data\") pod \"keystone-cron-29556601-msl6s\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.303694 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hks5\" (UniqueName: \"kubernetes.io/projected/2d23fe5e-07c3-4884-808f-adc6e0e8914f-kube-api-access-8hks5\") pod \"keystone-cron-29556601-msl6s\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.406066 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-config-data\") pod \"keystone-cron-29556601-msl6s\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.406137 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hks5\" (UniqueName: \"kubernetes.io/projected/2d23fe5e-07c3-4884-808f-adc6e0e8914f-kube-api-access-8hks5\") pod \"keystone-cron-29556601-msl6s\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.406228 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-combined-ca-bundle\") pod \"keystone-cron-29556601-msl6s\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.406305 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-fernet-keys\") pod \"keystone-cron-29556601-msl6s\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.413314 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-config-data\") pod \"keystone-cron-29556601-msl6s\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.413684 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-fernet-keys\") pod \"keystone-cron-29556601-msl6s\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.423960 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-combined-ca-bundle\") pod \"keystone-cron-29556601-msl6s\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.425771 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hks5\" (UniqueName: \"kubernetes.io/projected/2d23fe5e-07c3-4884-808f-adc6e0e8914f-kube-api-access-8hks5\") pod \"keystone-cron-29556601-msl6s\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:00 crc kubenswrapper[4930]: I0313 10:01:00.509156 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:01 crc kubenswrapper[4930]: I0313 10:01:01.025378 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29556601-msl6s"] Mar 13 10:01:01 crc kubenswrapper[4930]: I0313 10:01:01.544884 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556601-msl6s" event={"ID":"2d23fe5e-07c3-4884-808f-adc6e0e8914f","Type":"ContainerStarted","Data":"3739963f092b7dd40f691fa009b067900ec5770b199ee80e72814b975988daaf"} Mar 13 10:01:01 crc kubenswrapper[4930]: I0313 10:01:01.545400 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556601-msl6s" event={"ID":"2d23fe5e-07c3-4884-808f-adc6e0e8914f","Type":"ContainerStarted","Data":"e237c73608758b8688ae1e087aabb3966e49c4bc6a11c4ad352b8d2d05c6b730"} Mar 13 10:01:01 crc kubenswrapper[4930]: I0313 10:01:01.561723 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29556601-msl6s" podStartSLOduration=1.561703879 podStartE2EDuration="1.561703879s" podCreationTimestamp="2026-03-13 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:01:01.560623212 +0000 UTC m=+2902.310537889" watchObservedRunningTime="2026-03-13 10:01:01.561703879 +0000 UTC m=+2902.311618556" Mar 13 10:01:04 crc kubenswrapper[4930]: I0313 10:01:04.582311 4930 generic.go:334] "Generic (PLEG): container finished" podID="2d23fe5e-07c3-4884-808f-adc6e0e8914f" containerID="3739963f092b7dd40f691fa009b067900ec5770b199ee80e72814b975988daaf" exitCode=0 Mar 13 10:01:04 crc kubenswrapper[4930]: I0313 10:01:04.582723 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556601-msl6s" event={"ID":"2d23fe5e-07c3-4884-808f-adc6e0e8914f","Type":"ContainerDied","Data":"3739963f092b7dd40f691fa009b067900ec5770b199ee80e72814b975988daaf"} Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.049408 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.175506 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hks5\" (UniqueName: \"kubernetes.io/projected/2d23fe5e-07c3-4884-808f-adc6e0e8914f-kube-api-access-8hks5\") pod \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.175881 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-combined-ca-bundle\") pod \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.176034 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-fernet-keys\") pod \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.176154 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-config-data\") pod \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\" (UID: \"2d23fe5e-07c3-4884-808f-adc6e0e8914f\") " Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.199911 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2d23fe5e-07c3-4884-808f-adc6e0e8914f" (UID: "2d23fe5e-07c3-4884-808f-adc6e0e8914f"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.199969 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d23fe5e-07c3-4884-808f-adc6e0e8914f-kube-api-access-8hks5" (OuterVolumeSpecName: "kube-api-access-8hks5") pod "2d23fe5e-07c3-4884-808f-adc6e0e8914f" (UID: "2d23fe5e-07c3-4884-808f-adc6e0e8914f"). InnerVolumeSpecName "kube-api-access-8hks5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.216823 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d23fe5e-07c3-4884-808f-adc6e0e8914f" (UID: "2d23fe5e-07c3-4884-808f-adc6e0e8914f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.245686 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-config-data" (OuterVolumeSpecName: "config-data") pod "2d23fe5e-07c3-4884-808f-adc6e0e8914f" (UID: "2d23fe5e-07c3-4884-808f-adc6e0e8914f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.279452 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.279485 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hks5\" (UniqueName: \"kubernetes.io/projected/2d23fe5e-07c3-4884-808f-adc6e0e8914f-kube-api-access-8hks5\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.279497 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.279505 4930 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2d23fe5e-07c3-4884-808f-adc6e0e8914f-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.605453 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29556601-msl6s" event={"ID":"2d23fe5e-07c3-4884-808f-adc6e0e8914f","Type":"ContainerDied","Data":"e237c73608758b8688ae1e087aabb3966e49c4bc6a11c4ad352b8d2d05c6b730"} Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.605812 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e237c73608758b8688ae1e087aabb3966e49c4bc6a11c4ad352b8d2d05c6b730" Mar 13 10:01:06 crc kubenswrapper[4930]: I0313 10:01:06.605592 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29556601-msl6s" Mar 13 10:01:12 crc kubenswrapper[4930]: I0313 10:01:12.308752 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:01:12 crc kubenswrapper[4930]: I0313 10:01:12.309354 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:01:34 crc kubenswrapper[4930]: I0313 10:01:34.909214 4930 generic.go:334] "Generic (PLEG): container finished" podID="9498f5ee-52d0-4a7b-81d5-dc4f9e54416b" containerID="5c30a4c82b788b21624f33089c3c898a02049354fb2e300e5beb696a071c9c74" exitCode=0 Mar 13 10:01:34 crc kubenswrapper[4930]: I0313 10:01:34.909254 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" event={"ID":"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b","Type":"ContainerDied","Data":"5c30a4c82b788b21624f33089c3c898a02049354fb2e300e5beb696a071c9c74"} Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.409678 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.589115 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s7lft\" (UniqueName: \"kubernetes.io/projected/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-kube-api-access-s7lft\") pod \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.589239 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-inventory\") pod \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.589336 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-telemetry-power-monitoring-combined-ca-bundle\") pod \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.589406 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ssh-key-openstack-edpm-ipam\") pod \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.589556 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-0\") pod \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.589643 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-1\") pod \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.589707 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-2\") pod \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\" (UID: \"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b\") " Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.596947 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-kube-api-access-s7lft" (OuterVolumeSpecName: "kube-api-access-s7lft") pod "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b" (UID: "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b"). InnerVolumeSpecName "kube-api-access-s7lft". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.598666 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b" (UID: "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.627099 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b" (UID: "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.628851 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b" (UID: "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.629980 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-inventory" (OuterVolumeSpecName: "inventory") pod "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b" (UID: "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.634018 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b" (UID: "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.636790 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b" (UID: "9498f5ee-52d0-4a7b-81d5-dc4f9e54416b"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.693800 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s7lft\" (UniqueName: \"kubernetes.io/projected/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-kube-api-access-s7lft\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.693867 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.693883 4930 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.693895 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.693908 4930 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.693919 4930 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.693933 4930 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/9498f5ee-52d0-4a7b-81d5-dc4f9e54416b-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.938111 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" event={"ID":"9498f5ee-52d0-4a7b-81d5-dc4f9e54416b","Type":"ContainerDied","Data":"8273ab3f6f4cfeea591381dd4341d0683e5c514322b4b77d973b1d3b6b0415cc"} Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.938189 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8273ab3f6f4cfeea591381dd4341d0683e5c514322b4b77d973b1d3b6b0415cc" Mar 13 10:01:36 crc kubenswrapper[4930]: I0313 10:01:36.938287 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-wsnsh" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.043792 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv"] Mar 13 10:01:37 crc kubenswrapper[4930]: E0313 10:01:37.044369 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9498f5ee-52d0-4a7b-81d5-dc4f9e54416b" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.044387 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="9498f5ee-52d0-4a7b-81d5-dc4f9e54416b" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Mar 13 10:01:37 crc kubenswrapper[4930]: E0313 10:01:37.044399 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d23fe5e-07c3-4884-808f-adc6e0e8914f" containerName="keystone-cron" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.044407 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d23fe5e-07c3-4884-808f-adc6e0e8914f" containerName="keystone-cron" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.044628 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d23fe5e-07c3-4884-808f-adc6e0e8914f" containerName="keystone-cron" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.044645 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="9498f5ee-52d0-4a7b-81d5-dc4f9e54416b" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.045456 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.048489 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.049214 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.049594 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-94vzq" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.049880 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.051910 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.070789 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv"] Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.104410 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4swhr\" (UniqueName: \"kubernetes.io/projected/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-kube-api-access-4swhr\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.104613 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.104746 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.105247 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-ssh-key-openstack-edpm-ipam\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.107500 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.210356 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.210423 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.210529 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-ssh-key-openstack-edpm-ipam\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.210565 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.210743 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4swhr\" (UniqueName: \"kubernetes.io/projected/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-kube-api-access-4swhr\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.216397 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.217020 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.217178 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.217356 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-ssh-key-openstack-edpm-ipam\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.231172 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4swhr\" (UniqueName: \"kubernetes.io/projected/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-kube-api-access-4swhr\") pod \"logging-edpm-deployment-openstack-edpm-ipam-99kpv\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.409924 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.992023 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:01:37 crc kubenswrapper[4930]: I0313 10:01:37.992791 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv"] Mar 13 10:01:38 crc kubenswrapper[4930]: I0313 10:01:38.963337 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" event={"ID":"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2","Type":"ContainerStarted","Data":"41b21a8e4ecf8e6e9362e3e2623484ebc3e774929d059cae554e63847d0a3183"} Mar 13 10:01:38 crc kubenswrapper[4930]: I0313 10:01:38.963733 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" event={"ID":"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2","Type":"ContainerStarted","Data":"dfc2b0d9be718ab0a388f39467cb92d6b51cb00eed3a0a2c04ab8563f9275bdc"} Mar 13 10:01:42 crc kubenswrapper[4930]: I0313 10:01:42.308776 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:01:42 crc kubenswrapper[4930]: I0313 10:01:42.309462 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:01:42 crc kubenswrapper[4930]: I0313 10:01:42.309516 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 10:01:42 crc kubenswrapper[4930]: I0313 10:01:42.310475 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:01:42 crc kubenswrapper[4930]: I0313 10:01:42.310544 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" gracePeriod=600 Mar 13 10:01:42 crc kubenswrapper[4930]: E0313 10:01:42.453197 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:01:43 crc kubenswrapper[4930]: I0313 10:01:43.040980 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" exitCode=0 Mar 13 10:01:43 crc kubenswrapper[4930]: I0313 10:01:43.041028 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38"} Mar 13 10:01:43 crc kubenswrapper[4930]: I0313 10:01:43.041060 4930 scope.go:117] "RemoveContainer" containerID="5fdfdf4f6031ee1e4c460c1631bfff6cd403d76eae7f5fa87795d52154b3db24" Mar 13 10:01:43 crc kubenswrapper[4930]: I0313 10:01:43.041859 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:01:43 crc kubenswrapper[4930]: E0313 10:01:43.042120 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:01:43 crc kubenswrapper[4930]: I0313 10:01:43.123912 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" podStartSLOduration=5.5901628930000005 podStartE2EDuration="6.123888444s" podCreationTimestamp="2026-03-13 10:01:37 +0000 UTC" firstStartedPulling="2026-03-13 10:01:37.99165978 +0000 UTC m=+2938.741574467" lastFinishedPulling="2026-03-13 10:01:38.525385341 +0000 UTC m=+2939.275300018" observedRunningTime="2026-03-13 10:01:38.991471465 +0000 UTC m=+2939.741386142" watchObservedRunningTime="2026-03-13 10:01:43.123888444 +0000 UTC m=+2943.873803161" Mar 13 10:01:52 crc kubenswrapper[4930]: I0313 10:01:52.132976 4930 generic.go:334] "Generic (PLEG): container finished" podID="da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2" containerID="41b21a8e4ecf8e6e9362e3e2623484ebc3e774929d059cae554e63847d0a3183" exitCode=0 Mar 13 10:01:52 crc kubenswrapper[4930]: I0313 10:01:52.133045 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" event={"ID":"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2","Type":"ContainerDied","Data":"41b21a8e4ecf8e6e9362e3e2623484ebc3e774929d059cae554e63847d0a3183"} Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.622373 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.795418 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-logging-compute-config-data-1\") pod \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.795559 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-ssh-key-openstack-edpm-ipam\") pod \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.795641 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-inventory\") pod \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.795755 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-logging-compute-config-data-0\") pod \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.795835 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4swhr\" (UniqueName: \"kubernetes.io/projected/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-kube-api-access-4swhr\") pod \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\" (UID: \"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2\") " Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.805096 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-kube-api-access-4swhr" (OuterVolumeSpecName: "kube-api-access-4swhr") pod "da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2" (UID: "da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2"). InnerVolumeSpecName "kube-api-access-4swhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.829107 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2" (UID: "da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.830936 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2" (UID: "da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.833602 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-inventory" (OuterVolumeSpecName: "inventory") pod "da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2" (UID: "da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.833635 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2" (UID: "da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.901019 4930 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.901057 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.901069 4930 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-inventory\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.901079 4930 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:53 crc kubenswrapper[4930]: I0313 10:01:53.901089 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4swhr\" (UniqueName: \"kubernetes.io/projected/da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2-kube-api-access-4swhr\") on node \"crc\" DevicePath \"\"" Mar 13 10:01:54 crc kubenswrapper[4930]: I0313 10:01:54.161883 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" event={"ID":"da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2","Type":"ContainerDied","Data":"dfc2b0d9be718ab0a388f39467cb92d6b51cb00eed3a0a2c04ab8563f9275bdc"} Mar 13 10:01:54 crc kubenswrapper[4930]: I0313 10:01:54.161933 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfc2b0d9be718ab0a388f39467cb92d6b51cb00eed3a0a2c04ab8563f9275bdc" Mar 13 10:01:54 crc kubenswrapper[4930]: I0313 10:01:54.162000 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-99kpv" Mar 13 10:01:54 crc kubenswrapper[4930]: I0313 10:01:54.971003 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:01:54 crc kubenswrapper[4930]: E0313 10:01:54.971646 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:02:00 crc kubenswrapper[4930]: I0313 10:02:00.146848 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556602-lhlcv"] Mar 13 10:02:00 crc kubenswrapper[4930]: E0313 10:02:00.149255 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2" containerName="logging-edpm-deployment-openstack-edpm-ipam" Mar 13 10:02:00 crc kubenswrapper[4930]: I0313 10:02:00.149270 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2" containerName="logging-edpm-deployment-openstack-edpm-ipam" Mar 13 10:02:00 crc kubenswrapper[4930]: I0313 10:02:00.149612 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="da50c1ad-44ed-4c48-a7d1-d4e09cb1bed2" containerName="logging-edpm-deployment-openstack-edpm-ipam" Mar 13 10:02:00 crc kubenswrapper[4930]: I0313 10:02:00.150668 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556602-lhlcv" Mar 13 10:02:00 crc kubenswrapper[4930]: I0313 10:02:00.153336 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:02:00 crc kubenswrapper[4930]: I0313 10:02:00.153664 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:02:00 crc kubenswrapper[4930]: I0313 10:02:00.153911 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:02:00 crc kubenswrapper[4930]: I0313 10:02:00.161256 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556602-lhlcv"] Mar 13 10:02:00 crc kubenswrapper[4930]: I0313 10:02:00.171469 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhr6m\" (UniqueName: \"kubernetes.io/projected/e112b6bc-29f1-407f-b324-c5f0ec7917b2-kube-api-access-hhr6m\") pod \"auto-csr-approver-29556602-lhlcv\" (UID: \"e112b6bc-29f1-407f-b324-c5f0ec7917b2\") " pod="openshift-infra/auto-csr-approver-29556602-lhlcv" Mar 13 10:02:00 crc kubenswrapper[4930]: I0313 10:02:00.273924 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhr6m\" (UniqueName: \"kubernetes.io/projected/e112b6bc-29f1-407f-b324-c5f0ec7917b2-kube-api-access-hhr6m\") pod \"auto-csr-approver-29556602-lhlcv\" (UID: \"e112b6bc-29f1-407f-b324-c5f0ec7917b2\") " pod="openshift-infra/auto-csr-approver-29556602-lhlcv" Mar 13 10:02:00 crc kubenswrapper[4930]: I0313 10:02:00.306904 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhr6m\" (UniqueName: \"kubernetes.io/projected/e112b6bc-29f1-407f-b324-c5f0ec7917b2-kube-api-access-hhr6m\") pod \"auto-csr-approver-29556602-lhlcv\" (UID: \"e112b6bc-29f1-407f-b324-c5f0ec7917b2\") " pod="openshift-infra/auto-csr-approver-29556602-lhlcv" Mar 13 10:02:00 crc kubenswrapper[4930]: I0313 10:02:00.476039 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556602-lhlcv" Mar 13 10:02:01 crc kubenswrapper[4930]: I0313 10:02:01.122521 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556602-lhlcv"] Mar 13 10:02:01 crc kubenswrapper[4930]: I0313 10:02:01.252480 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556602-lhlcv" event={"ID":"e112b6bc-29f1-407f-b324-c5f0ec7917b2","Type":"ContainerStarted","Data":"7086f25bdc13cf20c4a477eba8c3e7243fcc67142e6a388d9b7bdc10386783bf"} Mar 13 10:02:03 crc kubenswrapper[4930]: I0313 10:02:03.276613 4930 generic.go:334] "Generic (PLEG): container finished" podID="e112b6bc-29f1-407f-b324-c5f0ec7917b2" containerID="007c0f5c70f64210b78ba9fc571927075f2e43025cc25f9e177d9cc437054cf5" exitCode=0 Mar 13 10:02:03 crc kubenswrapper[4930]: I0313 10:02:03.276660 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556602-lhlcv" event={"ID":"e112b6bc-29f1-407f-b324-c5f0ec7917b2","Type":"ContainerDied","Data":"007c0f5c70f64210b78ba9fc571927075f2e43025cc25f9e177d9cc437054cf5"} Mar 13 10:02:04 crc kubenswrapper[4930]: I0313 10:02:04.875215 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556602-lhlcv" Mar 13 10:02:05 crc kubenswrapper[4930]: I0313 10:02:05.029404 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhr6m\" (UniqueName: \"kubernetes.io/projected/e112b6bc-29f1-407f-b324-c5f0ec7917b2-kube-api-access-hhr6m\") pod \"e112b6bc-29f1-407f-b324-c5f0ec7917b2\" (UID: \"e112b6bc-29f1-407f-b324-c5f0ec7917b2\") " Mar 13 10:02:05 crc kubenswrapper[4930]: I0313 10:02:05.038343 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e112b6bc-29f1-407f-b324-c5f0ec7917b2-kube-api-access-hhr6m" (OuterVolumeSpecName: "kube-api-access-hhr6m") pod "e112b6bc-29f1-407f-b324-c5f0ec7917b2" (UID: "e112b6bc-29f1-407f-b324-c5f0ec7917b2"). InnerVolumeSpecName "kube-api-access-hhr6m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:02:05 crc kubenswrapper[4930]: I0313 10:02:05.132928 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhr6m\" (UniqueName: \"kubernetes.io/projected/e112b6bc-29f1-407f-b324-c5f0ec7917b2-kube-api-access-hhr6m\") on node \"crc\" DevicePath \"\"" Mar 13 10:02:05 crc kubenswrapper[4930]: I0313 10:02:05.300319 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556602-lhlcv" event={"ID":"e112b6bc-29f1-407f-b324-c5f0ec7917b2","Type":"ContainerDied","Data":"7086f25bdc13cf20c4a477eba8c3e7243fcc67142e6a388d9b7bdc10386783bf"} Mar 13 10:02:05 crc kubenswrapper[4930]: I0313 10:02:05.300356 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7086f25bdc13cf20c4a477eba8c3e7243fcc67142e6a388d9b7bdc10386783bf" Mar 13 10:02:05 crc kubenswrapper[4930]: I0313 10:02:05.300409 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556602-lhlcv" Mar 13 10:02:05 crc kubenswrapper[4930]: I0313 10:02:05.966469 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556596-tz96h"] Mar 13 10:02:05 crc kubenswrapper[4930]: I0313 10:02:05.971176 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:02:05 crc kubenswrapper[4930]: E0313 10:02:05.971555 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:02:05 crc kubenswrapper[4930]: I0313 10:02:05.987594 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556596-tz96h"] Mar 13 10:02:07 crc kubenswrapper[4930]: I0313 10:02:07.986520 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="597b92ee-ba9b-4e35-b135-4eb65be5a6f3" path="/var/lib/kubelet/pods/597b92ee-ba9b-4e35-b135-4eb65be5a6f3/volumes" Mar 13 10:02:20 crc kubenswrapper[4930]: I0313 10:02:20.971457 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:02:20 crc kubenswrapper[4930]: E0313 10:02:20.972726 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:02:31 crc kubenswrapper[4930]: I0313 10:02:31.687296 4930 scope.go:117] "RemoveContainer" containerID="0a90d80cd299121a5060eb9418b2347309a4e4fdcae71860d7a2668580158a5d" Mar 13 10:02:33 crc kubenswrapper[4930]: I0313 10:02:33.970990 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:02:33 crc kubenswrapper[4930]: E0313 10:02:33.972966 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:02:47 crc kubenswrapper[4930]: I0313 10:02:47.971622 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:02:47 crc kubenswrapper[4930]: E0313 10:02:47.972729 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:02:52 crc kubenswrapper[4930]: E0313 10:02:52.260824 4930 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.238:37094->38.102.83.238:33987: write tcp 38.102.83.238:37094->38.102.83.238:33987: write: broken pipe Mar 13 10:02:58 crc kubenswrapper[4930]: I0313 10:02:58.970910 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:02:58 crc kubenswrapper[4930]: E0313 10:02:58.971630 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:03:10 crc kubenswrapper[4930]: I0313 10:03:10.971020 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:03:10 crc kubenswrapper[4930]: E0313 10:03:10.971864 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:03:21 crc kubenswrapper[4930]: I0313 10:03:21.971063 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:03:21 crc kubenswrapper[4930]: E0313 10:03:21.972026 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:03:35 crc kubenswrapper[4930]: I0313 10:03:35.971189 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:03:35 crc kubenswrapper[4930]: E0313 10:03:35.971992 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:03:46 crc kubenswrapper[4930]: I0313 10:03:46.971574 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:03:46 crc kubenswrapper[4930]: E0313 10:03:46.972462 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.154860 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556604-k5xm4"] Mar 13 10:04:00 crc kubenswrapper[4930]: E0313 10:04:00.157731 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e112b6bc-29f1-407f-b324-c5f0ec7917b2" containerName="oc" Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.157769 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="e112b6bc-29f1-407f-b324-c5f0ec7917b2" containerName="oc" Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.158097 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="e112b6bc-29f1-407f-b324-c5f0ec7917b2" containerName="oc" Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.159345 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556604-k5xm4" Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.166338 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.167716 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.168906 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.185006 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556604-k5xm4"] Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.250501 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gj4xr\" (UniqueName: \"kubernetes.io/projected/ca1762ed-5f30-42a1-b549-ad60e97dfa8d-kube-api-access-gj4xr\") pod \"auto-csr-approver-29556604-k5xm4\" (UID: \"ca1762ed-5f30-42a1-b549-ad60e97dfa8d\") " pod="openshift-infra/auto-csr-approver-29556604-k5xm4" Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.353945 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gj4xr\" (UniqueName: \"kubernetes.io/projected/ca1762ed-5f30-42a1-b549-ad60e97dfa8d-kube-api-access-gj4xr\") pod \"auto-csr-approver-29556604-k5xm4\" (UID: \"ca1762ed-5f30-42a1-b549-ad60e97dfa8d\") " pod="openshift-infra/auto-csr-approver-29556604-k5xm4" Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.375140 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gj4xr\" (UniqueName: \"kubernetes.io/projected/ca1762ed-5f30-42a1-b549-ad60e97dfa8d-kube-api-access-gj4xr\") pod \"auto-csr-approver-29556604-k5xm4\" (UID: \"ca1762ed-5f30-42a1-b549-ad60e97dfa8d\") " pod="openshift-infra/auto-csr-approver-29556604-k5xm4" Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.486760 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556604-k5xm4" Mar 13 10:04:00 crc kubenswrapper[4930]: I0313 10:04:00.971969 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:04:00 crc kubenswrapper[4930]: E0313 10:04:00.972953 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:04:01 crc kubenswrapper[4930]: I0313 10:04:01.154511 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556604-k5xm4"] Mar 13 10:04:01 crc kubenswrapper[4930]: I0313 10:04:01.663474 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556604-k5xm4" event={"ID":"ca1762ed-5f30-42a1-b549-ad60e97dfa8d","Type":"ContainerStarted","Data":"12316d1c31b0f916afa2505625deace583608cc87dc5aba89a3909cca0f22641"} Mar 13 10:04:02 crc kubenswrapper[4930]: I0313 10:04:02.675688 4930 generic.go:334] "Generic (PLEG): container finished" podID="ca1762ed-5f30-42a1-b549-ad60e97dfa8d" containerID="6dec1d46b82212ef510faaa384c55d92199cb59edc4a7bae0fb9bf4aba9be3af" exitCode=0 Mar 13 10:04:02 crc kubenswrapper[4930]: I0313 10:04:02.675743 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556604-k5xm4" event={"ID":"ca1762ed-5f30-42a1-b549-ad60e97dfa8d","Type":"ContainerDied","Data":"6dec1d46b82212ef510faaa384c55d92199cb59edc4a7bae0fb9bf4aba9be3af"} Mar 13 10:04:04 crc kubenswrapper[4930]: I0313 10:04:04.090741 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556604-k5xm4" Mar 13 10:04:04 crc kubenswrapper[4930]: I0313 10:04:04.197622 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gj4xr\" (UniqueName: \"kubernetes.io/projected/ca1762ed-5f30-42a1-b549-ad60e97dfa8d-kube-api-access-gj4xr\") pod \"ca1762ed-5f30-42a1-b549-ad60e97dfa8d\" (UID: \"ca1762ed-5f30-42a1-b549-ad60e97dfa8d\") " Mar 13 10:04:04 crc kubenswrapper[4930]: I0313 10:04:04.203467 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca1762ed-5f30-42a1-b549-ad60e97dfa8d-kube-api-access-gj4xr" (OuterVolumeSpecName: "kube-api-access-gj4xr") pod "ca1762ed-5f30-42a1-b549-ad60e97dfa8d" (UID: "ca1762ed-5f30-42a1-b549-ad60e97dfa8d"). InnerVolumeSpecName "kube-api-access-gj4xr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:04:04 crc kubenswrapper[4930]: I0313 10:04:04.303270 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gj4xr\" (UniqueName: \"kubernetes.io/projected/ca1762ed-5f30-42a1-b549-ad60e97dfa8d-kube-api-access-gj4xr\") on node \"crc\" DevicePath \"\"" Mar 13 10:04:04 crc kubenswrapper[4930]: I0313 10:04:04.696769 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556604-k5xm4" event={"ID":"ca1762ed-5f30-42a1-b549-ad60e97dfa8d","Type":"ContainerDied","Data":"12316d1c31b0f916afa2505625deace583608cc87dc5aba89a3909cca0f22641"} Mar 13 10:04:04 crc kubenswrapper[4930]: I0313 10:04:04.696809 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="12316d1c31b0f916afa2505625deace583608cc87dc5aba89a3909cca0f22641" Mar 13 10:04:04 crc kubenswrapper[4930]: I0313 10:04:04.696895 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556604-k5xm4" Mar 13 10:04:05 crc kubenswrapper[4930]: I0313 10:04:05.186834 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556598-td629"] Mar 13 10:04:05 crc kubenswrapper[4930]: I0313 10:04:05.202631 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556598-td629"] Mar 13 10:04:05 crc kubenswrapper[4930]: I0313 10:04:05.995788 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b38a35aa-d00e-4cff-960c-c933212cdeb4" path="/var/lib/kubelet/pods/b38a35aa-d00e-4cff-960c-c933212cdeb4/volumes" Mar 13 10:04:15 crc kubenswrapper[4930]: I0313 10:04:15.978746 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:04:15 crc kubenswrapper[4930]: E0313 10:04:15.979983 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:04:26 crc kubenswrapper[4930]: I0313 10:04:26.971751 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:04:26 crc kubenswrapper[4930]: E0313 10:04:26.972716 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:04:31 crc kubenswrapper[4930]: I0313 10:04:31.798107 4930 scope.go:117] "RemoveContainer" containerID="71a0d6451cc902c4c2a78587b5efdfa26f29ecd49ce737d016a1040341cae805" Mar 13 10:04:41 crc kubenswrapper[4930]: I0313 10:04:41.971078 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:04:41 crc kubenswrapper[4930]: E0313 10:04:41.971946 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.066057 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-hmdsk"] Mar 13 10:04:47 crc kubenswrapper[4930]: E0313 10:04:47.067312 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca1762ed-5f30-42a1-b549-ad60e97dfa8d" containerName="oc" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.067334 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca1762ed-5f30-42a1-b549-ad60e97dfa8d" containerName="oc" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.067652 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca1762ed-5f30-42a1-b549-ad60e97dfa8d" containerName="oc" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.069943 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.082818 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hmdsk"] Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.139149 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znv2g\" (UniqueName: \"kubernetes.io/projected/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-kube-api-access-znv2g\") pod \"redhat-operators-hmdsk\" (UID: \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\") " pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.139207 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-catalog-content\") pod \"redhat-operators-hmdsk\" (UID: \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\") " pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.139306 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-utilities\") pod \"redhat-operators-hmdsk\" (UID: \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\") " pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.242126 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znv2g\" (UniqueName: \"kubernetes.io/projected/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-kube-api-access-znv2g\") pod \"redhat-operators-hmdsk\" (UID: \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\") " pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.242181 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-catalog-content\") pod \"redhat-operators-hmdsk\" (UID: \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\") " pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.242275 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-utilities\") pod \"redhat-operators-hmdsk\" (UID: \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\") " pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.242938 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-catalog-content\") pod \"redhat-operators-hmdsk\" (UID: \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\") " pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.242992 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-utilities\") pod \"redhat-operators-hmdsk\" (UID: \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\") " pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.264756 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znv2g\" (UniqueName: \"kubernetes.io/projected/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-kube-api-access-znv2g\") pod \"redhat-operators-hmdsk\" (UID: \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\") " pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:04:47 crc kubenswrapper[4930]: I0313 10:04:47.431484 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:04:48 crc kubenswrapper[4930]: I0313 10:04:48.008529 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-hmdsk"] Mar 13 10:04:48 crc kubenswrapper[4930]: I0313 10:04:48.258655 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmdsk" event={"ID":"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3","Type":"ContainerStarted","Data":"c771d37b9eeaa94a03a661d92f27a0c9da083a459645ed17e68efd39522b7c15"} Mar 13 10:04:49 crc kubenswrapper[4930]: I0313 10:04:49.270905 4930 generic.go:334] "Generic (PLEG): container finished" podID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerID="c9e6423cfc1111ea727688fb4bcf5ac22960f1d36fe4f1170b1b6f244962a3e9" exitCode=0 Mar 13 10:04:49 crc kubenswrapper[4930]: I0313 10:04:49.271005 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmdsk" event={"ID":"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3","Type":"ContainerDied","Data":"c9e6423cfc1111ea727688fb4bcf5ac22960f1d36fe4f1170b1b6f244962a3e9"} Mar 13 10:04:51 crc kubenswrapper[4930]: I0313 10:04:51.300052 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmdsk" event={"ID":"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3","Type":"ContainerStarted","Data":"c61ed12e1ecebae4b0edf0c24d99fc727c6dc887dcd6dd93cf3e12d543fc8127"} Mar 13 10:04:54 crc kubenswrapper[4930]: I0313 10:04:54.972051 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:04:54 crc kubenswrapper[4930]: E0313 10:04:54.974552 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:04:58 crc kubenswrapper[4930]: I0313 10:04:58.378151 4930 generic.go:334] "Generic (PLEG): container finished" podID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerID="c61ed12e1ecebae4b0edf0c24d99fc727c6dc887dcd6dd93cf3e12d543fc8127" exitCode=0 Mar 13 10:04:58 crc kubenswrapper[4930]: I0313 10:04:58.378204 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmdsk" event={"ID":"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3","Type":"ContainerDied","Data":"c61ed12e1ecebae4b0edf0c24d99fc727c6dc887dcd6dd93cf3e12d543fc8127"} Mar 13 10:04:59 crc kubenswrapper[4930]: I0313 10:04:59.394006 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmdsk" event={"ID":"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3","Type":"ContainerStarted","Data":"f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737"} Mar 13 10:04:59 crc kubenswrapper[4930]: I0313 10:04:59.411935 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-hmdsk" podStartSLOduration=2.664809696 podStartE2EDuration="12.41191705s" podCreationTimestamp="2026-03-13 10:04:47 +0000 UTC" firstStartedPulling="2026-03-13 10:04:49.273364514 +0000 UTC m=+3130.023279191" lastFinishedPulling="2026-03-13 10:04:59.020471868 +0000 UTC m=+3139.770386545" observedRunningTime="2026-03-13 10:04:59.409529571 +0000 UTC m=+3140.159444248" watchObservedRunningTime="2026-03-13 10:04:59.41191705 +0000 UTC m=+3140.161831727" Mar 13 10:05:07 crc kubenswrapper[4930]: I0313 10:05:07.432369 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:05:07 crc kubenswrapper[4930]: I0313 10:05:07.432983 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:05:07 crc kubenswrapper[4930]: I0313 10:05:07.971237 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:05:07 crc kubenswrapper[4930]: E0313 10:05:07.971872 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:05:08 crc kubenswrapper[4930]: I0313 10:05:08.484873 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hmdsk" podUID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerName="registry-server" probeResult="failure" output=< Mar 13 10:05:08 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:05:08 crc kubenswrapper[4930]: > Mar 13 10:05:18 crc kubenswrapper[4930]: I0313 10:05:18.502727 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hmdsk" podUID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerName="registry-server" probeResult="failure" output=< Mar 13 10:05:18 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:05:18 crc kubenswrapper[4930]: > Mar 13 10:05:19 crc kubenswrapper[4930]: I0313 10:05:19.979838 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:05:19 crc kubenswrapper[4930]: E0313 10:05:19.980426 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:05:28 crc kubenswrapper[4930]: I0313 10:05:28.487612 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-hmdsk" podUID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerName="registry-server" probeResult="failure" output=< Mar 13 10:05:28 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:05:28 crc kubenswrapper[4930]: > Mar 13 10:05:32 crc kubenswrapper[4930]: I0313 10:05:32.971088 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:05:32 crc kubenswrapper[4930]: E0313 10:05:32.971880 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:05:37 crc kubenswrapper[4930]: I0313 10:05:37.499707 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:05:37 crc kubenswrapper[4930]: I0313 10:05:37.562316 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:05:37 crc kubenswrapper[4930]: I0313 10:05:37.752910 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hmdsk"] Mar 13 10:05:38 crc kubenswrapper[4930]: I0313 10:05:38.811848 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-hmdsk" podUID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerName="registry-server" containerID="cri-o://f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737" gracePeriod=2 Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.433966 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.528756 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-catalog-content\") pod \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\" (UID: \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\") " Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.529033 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-utilities\") pod \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\" (UID: \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\") " Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.529211 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znv2g\" (UniqueName: \"kubernetes.io/projected/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-kube-api-access-znv2g\") pod \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\" (UID: \"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3\") " Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.529885 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-utilities" (OuterVolumeSpecName: "utilities") pod "19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" (UID: "19a7bb07-f2c1-450c-94ae-cf4deecc4dc3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.530209 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.538856 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-kube-api-access-znv2g" (OuterVolumeSpecName: "kube-api-access-znv2g") pod "19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" (UID: "19a7bb07-f2c1-450c-94ae-cf4deecc4dc3"). InnerVolumeSpecName "kube-api-access-znv2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.632893 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znv2g\" (UniqueName: \"kubernetes.io/projected/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-kube-api-access-znv2g\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.670058 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" (UID: "19a7bb07-f2c1-450c-94ae-cf4deecc4dc3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.735763 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.824238 4930 generic.go:334] "Generic (PLEG): container finished" podID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerID="f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737" exitCode=0 Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.824291 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmdsk" event={"ID":"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3","Type":"ContainerDied","Data":"f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737"} Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.824323 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-hmdsk" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.824345 4930 scope.go:117] "RemoveContainer" containerID="f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.824329 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-hmdsk" event={"ID":"19a7bb07-f2c1-450c-94ae-cf4deecc4dc3","Type":"ContainerDied","Data":"c771d37b9eeaa94a03a661d92f27a0c9da083a459645ed17e68efd39522b7c15"} Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.860367 4930 scope.go:117] "RemoveContainer" containerID="c61ed12e1ecebae4b0edf0c24d99fc727c6dc887dcd6dd93cf3e12d543fc8127" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.860933 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-hmdsk"] Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.871930 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-hmdsk"] Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.899199 4930 scope.go:117] "RemoveContainer" containerID="c9e6423cfc1111ea727688fb4bcf5ac22960f1d36fe4f1170b1b6f244962a3e9" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.957521 4930 scope.go:117] "RemoveContainer" containerID="f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737" Mar 13 10:05:39 crc kubenswrapper[4930]: E0313 10:05:39.957988 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737\": container with ID starting with f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737 not found: ID does not exist" containerID="f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.958035 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737"} err="failed to get container status \"f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737\": rpc error: code = NotFound desc = could not find container \"f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737\": container with ID starting with f89822044ea75325094cc3317de732ab3dd0d31d9607ec06a7a1c041438c9737 not found: ID does not exist" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.958061 4930 scope.go:117] "RemoveContainer" containerID="c61ed12e1ecebae4b0edf0c24d99fc727c6dc887dcd6dd93cf3e12d543fc8127" Mar 13 10:05:39 crc kubenswrapper[4930]: E0313 10:05:39.958662 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c61ed12e1ecebae4b0edf0c24d99fc727c6dc887dcd6dd93cf3e12d543fc8127\": container with ID starting with c61ed12e1ecebae4b0edf0c24d99fc727c6dc887dcd6dd93cf3e12d543fc8127 not found: ID does not exist" containerID="c61ed12e1ecebae4b0edf0c24d99fc727c6dc887dcd6dd93cf3e12d543fc8127" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.958701 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c61ed12e1ecebae4b0edf0c24d99fc727c6dc887dcd6dd93cf3e12d543fc8127"} err="failed to get container status \"c61ed12e1ecebae4b0edf0c24d99fc727c6dc887dcd6dd93cf3e12d543fc8127\": rpc error: code = NotFound desc = could not find container \"c61ed12e1ecebae4b0edf0c24d99fc727c6dc887dcd6dd93cf3e12d543fc8127\": container with ID starting with c61ed12e1ecebae4b0edf0c24d99fc727c6dc887dcd6dd93cf3e12d543fc8127 not found: ID does not exist" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.958716 4930 scope.go:117] "RemoveContainer" containerID="c9e6423cfc1111ea727688fb4bcf5ac22960f1d36fe4f1170b1b6f244962a3e9" Mar 13 10:05:39 crc kubenswrapper[4930]: E0313 10:05:39.958956 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9e6423cfc1111ea727688fb4bcf5ac22960f1d36fe4f1170b1b6f244962a3e9\": container with ID starting with c9e6423cfc1111ea727688fb4bcf5ac22960f1d36fe4f1170b1b6f244962a3e9 not found: ID does not exist" containerID="c9e6423cfc1111ea727688fb4bcf5ac22960f1d36fe4f1170b1b6f244962a3e9" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.959064 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9e6423cfc1111ea727688fb4bcf5ac22960f1d36fe4f1170b1b6f244962a3e9"} err="failed to get container status \"c9e6423cfc1111ea727688fb4bcf5ac22960f1d36fe4f1170b1b6f244962a3e9\": rpc error: code = NotFound desc = could not find container \"c9e6423cfc1111ea727688fb4bcf5ac22960f1d36fe4f1170b1b6f244962a3e9\": container with ID starting with c9e6423cfc1111ea727688fb4bcf5ac22960f1d36fe4f1170b1b6f244962a3e9 not found: ID does not exist" Mar 13 10:05:39 crc kubenswrapper[4930]: I0313 10:05:39.993589 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" path="/var/lib/kubelet/pods/19a7bb07-f2c1-450c-94ae-cf4deecc4dc3/volumes" Mar 13 10:05:45 crc kubenswrapper[4930]: I0313 10:05:45.971706 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:05:45 crc kubenswrapper[4930]: E0313 10:05:45.972383 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:05:59 crc kubenswrapper[4930]: I0313 10:05:59.980263 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:05:59 crc kubenswrapper[4930]: E0313 10:05:59.981132 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.192212 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556606-klqjt"] Mar 13 10:06:00 crc kubenswrapper[4930]: E0313 10:06:00.192792 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerName="extract-content" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.192810 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerName="extract-content" Mar 13 10:06:00 crc kubenswrapper[4930]: E0313 10:06:00.192843 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerName="extract-utilities" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.192850 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerName="extract-utilities" Mar 13 10:06:00 crc kubenswrapper[4930]: E0313 10:06:00.192886 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerName="registry-server" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.192892 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerName="registry-server" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.193102 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="19a7bb07-f2c1-450c-94ae-cf4deecc4dc3" containerName="registry-server" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.194055 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556606-klqjt" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.196517 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.197617 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.198092 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.204775 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556606-klqjt"] Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.299937 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbv2z\" (UniqueName: \"kubernetes.io/projected/01c99807-49ba-4ee9-a14a-f746b6f00e62-kube-api-access-nbv2z\") pod \"auto-csr-approver-29556606-klqjt\" (UID: \"01c99807-49ba-4ee9-a14a-f746b6f00e62\") " pod="openshift-infra/auto-csr-approver-29556606-klqjt" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.402962 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbv2z\" (UniqueName: \"kubernetes.io/projected/01c99807-49ba-4ee9-a14a-f746b6f00e62-kube-api-access-nbv2z\") pod \"auto-csr-approver-29556606-klqjt\" (UID: \"01c99807-49ba-4ee9-a14a-f746b6f00e62\") " pod="openshift-infra/auto-csr-approver-29556606-klqjt" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.425226 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbv2z\" (UniqueName: \"kubernetes.io/projected/01c99807-49ba-4ee9-a14a-f746b6f00e62-kube-api-access-nbv2z\") pod \"auto-csr-approver-29556606-klqjt\" (UID: \"01c99807-49ba-4ee9-a14a-f746b6f00e62\") " pod="openshift-infra/auto-csr-approver-29556606-klqjt" Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.514315 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556606-klqjt" Mar 13 10:06:00 crc kubenswrapper[4930]: W0313 10:06:00.995778 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod01c99807_49ba_4ee9_a14a_f746b6f00e62.slice/crio-0a449545a081456a88941b026f761a64cdda420e6df739e1a403bbf4a891ce40 WatchSource:0}: Error finding container 0a449545a081456a88941b026f761a64cdda420e6df739e1a403bbf4a891ce40: Status 404 returned error can't find the container with id 0a449545a081456a88941b026f761a64cdda420e6df739e1a403bbf4a891ce40 Mar 13 10:06:00 crc kubenswrapper[4930]: I0313 10:06:00.998697 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556606-klqjt"] Mar 13 10:06:01 crc kubenswrapper[4930]: I0313 10:06:01.047084 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556606-klqjt" event={"ID":"01c99807-49ba-4ee9-a14a-f746b6f00e62","Type":"ContainerStarted","Data":"0a449545a081456a88941b026f761a64cdda420e6df739e1a403bbf4a891ce40"} Mar 13 10:06:04 crc kubenswrapper[4930]: I0313 10:06:04.081048 4930 generic.go:334] "Generic (PLEG): container finished" podID="01c99807-49ba-4ee9-a14a-f746b6f00e62" containerID="aa281c345ef0a3caa0b8dfc91b5d14b6c4ec6f0d3add501b2738d0cd1495273f" exitCode=0 Mar 13 10:06:04 crc kubenswrapper[4930]: I0313 10:06:04.081155 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556606-klqjt" event={"ID":"01c99807-49ba-4ee9-a14a-f746b6f00e62","Type":"ContainerDied","Data":"aa281c345ef0a3caa0b8dfc91b5d14b6c4ec6f0d3add501b2738d0cd1495273f"} Mar 13 10:06:05 crc kubenswrapper[4930]: I0313 10:06:05.533219 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556606-klqjt" Mar 13 10:06:05 crc kubenswrapper[4930]: I0313 10:06:05.732154 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nbv2z\" (UniqueName: \"kubernetes.io/projected/01c99807-49ba-4ee9-a14a-f746b6f00e62-kube-api-access-nbv2z\") pod \"01c99807-49ba-4ee9-a14a-f746b6f00e62\" (UID: \"01c99807-49ba-4ee9-a14a-f746b6f00e62\") " Mar 13 10:06:05 crc kubenswrapper[4930]: I0313 10:06:05.739723 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01c99807-49ba-4ee9-a14a-f746b6f00e62-kube-api-access-nbv2z" (OuterVolumeSpecName: "kube-api-access-nbv2z") pod "01c99807-49ba-4ee9-a14a-f746b6f00e62" (UID: "01c99807-49ba-4ee9-a14a-f746b6f00e62"). InnerVolumeSpecName "kube-api-access-nbv2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:06:05 crc kubenswrapper[4930]: I0313 10:06:05.835055 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nbv2z\" (UniqueName: \"kubernetes.io/projected/01c99807-49ba-4ee9-a14a-f746b6f00e62-kube-api-access-nbv2z\") on node \"crc\" DevicePath \"\"" Mar 13 10:06:06 crc kubenswrapper[4930]: I0313 10:06:06.103224 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556606-klqjt" event={"ID":"01c99807-49ba-4ee9-a14a-f746b6f00e62","Type":"ContainerDied","Data":"0a449545a081456a88941b026f761a64cdda420e6df739e1a403bbf4a891ce40"} Mar 13 10:06:06 crc kubenswrapper[4930]: I0313 10:06:06.103275 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a449545a081456a88941b026f761a64cdda420e6df739e1a403bbf4a891ce40" Mar 13 10:06:06 crc kubenswrapper[4930]: I0313 10:06:06.103282 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556606-klqjt" Mar 13 10:06:06 crc kubenswrapper[4930]: I0313 10:06:06.611399 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556600-9cv2j"] Mar 13 10:06:06 crc kubenswrapper[4930]: I0313 10:06:06.622703 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556600-9cv2j"] Mar 13 10:06:07 crc kubenswrapper[4930]: I0313 10:06:07.985418 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679" path="/var/lib/kubelet/pods/a2c3e9bb-ea9c-4cf5-a3c6-37d1e85b4679/volumes" Mar 13 10:06:10 crc kubenswrapper[4930]: I0313 10:06:10.971314 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:06:10 crc kubenswrapper[4930]: E0313 10:06:10.972778 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:06:23 crc kubenswrapper[4930]: I0313 10:06:23.975207 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:06:23 crc kubenswrapper[4930]: E0313 10:06:23.976007 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:06:31 crc kubenswrapper[4930]: I0313 10:06:31.930165 4930 scope.go:117] "RemoveContainer" containerID="b1571594ae7616433e672271e95e4c41cc8610001fa6780c29bcb98cdb1fa2e6" Mar 13 10:06:34 crc kubenswrapper[4930]: I0313 10:06:34.971765 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:06:34 crc kubenswrapper[4930]: E0313 10:06:34.974298 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:06:45 crc kubenswrapper[4930]: I0313 10:06:45.970898 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:06:46 crc kubenswrapper[4930]: I0313 10:06:46.550525 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"f728726a2be84d2953c1136412cbb0280b9a681f89803670792f59768fcb7bbe"} Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.583170 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-xts2v"] Mar 13 10:07:52 crc kubenswrapper[4930]: E0313 10:07:52.584353 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="01c99807-49ba-4ee9-a14a-f746b6f00e62" containerName="oc" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.584371 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="01c99807-49ba-4ee9-a14a-f746b6f00e62" containerName="oc" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.584674 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="01c99807-49ba-4ee9-a14a-f746b6f00e62" containerName="oc" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.586799 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.602939 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xts2v"] Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.685492 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/726723ef-06d8-4873-8d16-383c15165e23-catalog-content\") pod \"community-operators-xts2v\" (UID: \"726723ef-06d8-4873-8d16-383c15165e23\") " pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.685533 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh5wn\" (UniqueName: \"kubernetes.io/projected/726723ef-06d8-4873-8d16-383c15165e23-kube-api-access-sh5wn\") pod \"community-operators-xts2v\" (UID: \"726723ef-06d8-4873-8d16-383c15165e23\") " pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.685714 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/726723ef-06d8-4873-8d16-383c15165e23-utilities\") pod \"community-operators-xts2v\" (UID: \"726723ef-06d8-4873-8d16-383c15165e23\") " pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.788412 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/726723ef-06d8-4873-8d16-383c15165e23-catalog-content\") pod \"community-operators-xts2v\" (UID: \"726723ef-06d8-4873-8d16-383c15165e23\") " pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.788798 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh5wn\" (UniqueName: \"kubernetes.io/projected/726723ef-06d8-4873-8d16-383c15165e23-kube-api-access-sh5wn\") pod \"community-operators-xts2v\" (UID: \"726723ef-06d8-4873-8d16-383c15165e23\") " pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.788904 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/726723ef-06d8-4873-8d16-383c15165e23-utilities\") pod \"community-operators-xts2v\" (UID: \"726723ef-06d8-4873-8d16-383c15165e23\") " pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.790717 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/726723ef-06d8-4873-8d16-383c15165e23-catalog-content\") pod \"community-operators-xts2v\" (UID: \"726723ef-06d8-4873-8d16-383c15165e23\") " pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.791011 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/726723ef-06d8-4873-8d16-383c15165e23-utilities\") pod \"community-operators-xts2v\" (UID: \"726723ef-06d8-4873-8d16-383c15165e23\") " pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.813403 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh5wn\" (UniqueName: \"kubernetes.io/projected/726723ef-06d8-4873-8d16-383c15165e23-kube-api-access-sh5wn\") pod \"community-operators-xts2v\" (UID: \"726723ef-06d8-4873-8d16-383c15165e23\") " pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:07:52 crc kubenswrapper[4930]: I0313 10:07:52.908030 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:07:53 crc kubenswrapper[4930]: I0313 10:07:53.448974 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-xts2v"] Mar 13 10:07:54 crc kubenswrapper[4930]: I0313 10:07:54.351969 4930 generic.go:334] "Generic (PLEG): container finished" podID="726723ef-06d8-4873-8d16-383c15165e23" containerID="f5d2b506009650f27b291cdca51785954b17925291d6e1f2285a147a18b0ef90" exitCode=0 Mar 13 10:07:54 crc kubenswrapper[4930]: I0313 10:07:54.352033 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xts2v" event={"ID":"726723ef-06d8-4873-8d16-383c15165e23","Type":"ContainerDied","Data":"f5d2b506009650f27b291cdca51785954b17925291d6e1f2285a147a18b0ef90"} Mar 13 10:07:54 crc kubenswrapper[4930]: I0313 10:07:54.352478 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xts2v" event={"ID":"726723ef-06d8-4873-8d16-383c15165e23","Type":"ContainerStarted","Data":"aa382d292a3f0985e87ecc2f15ca35eb789c95597730915639f01775fde8f7a3"} Mar 13 10:07:54 crc kubenswrapper[4930]: I0313 10:07:54.354671 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:07:56 crc kubenswrapper[4930]: I0313 10:07:56.381871 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xts2v" event={"ID":"726723ef-06d8-4873-8d16-383c15165e23","Type":"ContainerStarted","Data":"3b6b7481d91fcf99ac7fe598561b0da4930c4763dc54bb3b29b4aed0be75511e"} Mar 13 10:07:58 crc kubenswrapper[4930]: I0313 10:07:58.406044 4930 generic.go:334] "Generic (PLEG): container finished" podID="726723ef-06d8-4873-8d16-383c15165e23" containerID="3b6b7481d91fcf99ac7fe598561b0da4930c4763dc54bb3b29b4aed0be75511e" exitCode=0 Mar 13 10:07:58 crc kubenswrapper[4930]: I0313 10:07:58.406131 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xts2v" event={"ID":"726723ef-06d8-4873-8d16-383c15165e23","Type":"ContainerDied","Data":"3b6b7481d91fcf99ac7fe598561b0da4930c4763dc54bb3b29b4aed0be75511e"} Mar 13 10:07:59 crc kubenswrapper[4930]: I0313 10:07:59.420896 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xts2v" event={"ID":"726723ef-06d8-4873-8d16-383c15165e23","Type":"ContainerStarted","Data":"e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362"} Mar 13 10:07:59 crc kubenswrapper[4930]: I0313 10:07:59.443366 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-xts2v" podStartSLOduration=2.752281381 podStartE2EDuration="7.44334686s" podCreationTimestamp="2026-03-13 10:07:52 +0000 UTC" firstStartedPulling="2026-03-13 10:07:54.354350308 +0000 UTC m=+3315.104264985" lastFinishedPulling="2026-03-13 10:07:59.045415787 +0000 UTC m=+3319.795330464" observedRunningTime="2026-03-13 10:07:59.441609647 +0000 UTC m=+3320.191524324" watchObservedRunningTime="2026-03-13 10:07:59.44334686 +0000 UTC m=+3320.193261537" Mar 13 10:08:00 crc kubenswrapper[4930]: I0313 10:08:00.156319 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556608-zczc8"] Mar 13 10:08:00 crc kubenswrapper[4930]: I0313 10:08:00.158011 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556608-zczc8" Mar 13 10:08:00 crc kubenswrapper[4930]: I0313 10:08:00.161025 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:08:00 crc kubenswrapper[4930]: I0313 10:08:00.161551 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:08:00 crc kubenswrapper[4930]: I0313 10:08:00.161564 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:08:00 crc kubenswrapper[4930]: I0313 10:08:00.214367 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556608-zczc8"] Mar 13 10:08:00 crc kubenswrapper[4930]: I0313 10:08:00.313156 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8m7f\" (UniqueName: \"kubernetes.io/projected/3d41547f-e823-475f-8495-ebda890df4a3-kube-api-access-f8m7f\") pod \"auto-csr-approver-29556608-zczc8\" (UID: \"3d41547f-e823-475f-8495-ebda890df4a3\") " pod="openshift-infra/auto-csr-approver-29556608-zczc8" Mar 13 10:08:00 crc kubenswrapper[4930]: I0313 10:08:00.416844 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8m7f\" (UniqueName: \"kubernetes.io/projected/3d41547f-e823-475f-8495-ebda890df4a3-kube-api-access-f8m7f\") pod \"auto-csr-approver-29556608-zczc8\" (UID: \"3d41547f-e823-475f-8495-ebda890df4a3\") " pod="openshift-infra/auto-csr-approver-29556608-zczc8" Mar 13 10:08:00 crc kubenswrapper[4930]: I0313 10:08:00.460559 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8m7f\" (UniqueName: \"kubernetes.io/projected/3d41547f-e823-475f-8495-ebda890df4a3-kube-api-access-f8m7f\") pod \"auto-csr-approver-29556608-zczc8\" (UID: \"3d41547f-e823-475f-8495-ebda890df4a3\") " pod="openshift-infra/auto-csr-approver-29556608-zczc8" Mar 13 10:08:00 crc kubenswrapper[4930]: I0313 10:08:00.518747 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556608-zczc8" Mar 13 10:08:01 crc kubenswrapper[4930]: I0313 10:08:01.104061 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556608-zczc8"] Mar 13 10:08:01 crc kubenswrapper[4930]: I0313 10:08:01.447672 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556608-zczc8" event={"ID":"3d41547f-e823-475f-8495-ebda890df4a3","Type":"ContainerStarted","Data":"b0184659fe91e144cbf83a033f3a63ae24c981e61c342434da5f51cbf8c9c223"} Mar 13 10:08:02 crc kubenswrapper[4930]: I0313 10:08:02.488508 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556608-zczc8" event={"ID":"3d41547f-e823-475f-8495-ebda890df4a3","Type":"ContainerStarted","Data":"b070893d9a53bfba06eea2e119897fe6ea9c1054bb9913730f6260e6cb98716d"} Mar 13 10:08:02 crc kubenswrapper[4930]: I0313 10:08:02.533381 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556608-zczc8" podStartSLOduration=1.838227352 podStartE2EDuration="2.533358832s" podCreationTimestamp="2026-03-13 10:08:00 +0000 UTC" firstStartedPulling="2026-03-13 10:08:01.128165968 +0000 UTC m=+3321.878080645" lastFinishedPulling="2026-03-13 10:08:01.823297448 +0000 UTC m=+3322.573212125" observedRunningTime="2026-03-13 10:08:02.510823506 +0000 UTC m=+3323.260738183" watchObservedRunningTime="2026-03-13 10:08:02.533358832 +0000 UTC m=+3323.283273509" Mar 13 10:08:02 crc kubenswrapper[4930]: I0313 10:08:02.908375 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:08:02 crc kubenswrapper[4930]: I0313 10:08:02.909269 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:08:02 crc kubenswrapper[4930]: I0313 10:08:02.961670 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:08:03 crc kubenswrapper[4930]: I0313 10:08:03.505840 4930 generic.go:334] "Generic (PLEG): container finished" podID="3d41547f-e823-475f-8495-ebda890df4a3" containerID="b070893d9a53bfba06eea2e119897fe6ea9c1054bb9913730f6260e6cb98716d" exitCode=0 Mar 13 10:08:03 crc kubenswrapper[4930]: I0313 10:08:03.509633 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556608-zczc8" event={"ID":"3d41547f-e823-475f-8495-ebda890df4a3","Type":"ContainerDied","Data":"b070893d9a53bfba06eea2e119897fe6ea9c1054bb9913730f6260e6cb98716d"} Mar 13 10:08:04 crc kubenswrapper[4930]: I0313 10:08:04.995742 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556608-zczc8" Mar 13 10:08:05 crc kubenswrapper[4930]: I0313 10:08:05.050116 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f8m7f\" (UniqueName: \"kubernetes.io/projected/3d41547f-e823-475f-8495-ebda890df4a3-kube-api-access-f8m7f\") pod \"3d41547f-e823-475f-8495-ebda890df4a3\" (UID: \"3d41547f-e823-475f-8495-ebda890df4a3\") " Mar 13 10:08:05 crc kubenswrapper[4930]: I0313 10:08:05.074010 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d41547f-e823-475f-8495-ebda890df4a3-kube-api-access-f8m7f" (OuterVolumeSpecName: "kube-api-access-f8m7f") pod "3d41547f-e823-475f-8495-ebda890df4a3" (UID: "3d41547f-e823-475f-8495-ebda890df4a3"). InnerVolumeSpecName "kube-api-access-f8m7f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:08:05 crc kubenswrapper[4930]: I0313 10:08:05.153947 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f8m7f\" (UniqueName: \"kubernetes.io/projected/3d41547f-e823-475f-8495-ebda890df4a3-kube-api-access-f8m7f\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:05 crc kubenswrapper[4930]: I0313 10:08:05.531383 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556608-zczc8" event={"ID":"3d41547f-e823-475f-8495-ebda890df4a3","Type":"ContainerDied","Data":"b0184659fe91e144cbf83a033f3a63ae24c981e61c342434da5f51cbf8c9c223"} Mar 13 10:08:05 crc kubenswrapper[4930]: I0313 10:08:05.531746 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0184659fe91e144cbf83a033f3a63ae24c981e61c342434da5f51cbf8c9c223" Mar 13 10:08:05 crc kubenswrapper[4930]: I0313 10:08:05.531454 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556608-zczc8" Mar 13 10:08:05 crc kubenswrapper[4930]: I0313 10:08:05.595734 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556602-lhlcv"] Mar 13 10:08:05 crc kubenswrapper[4930]: I0313 10:08:05.611821 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556602-lhlcv"] Mar 13 10:08:05 crc kubenswrapper[4930]: I0313 10:08:05.987773 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e112b6bc-29f1-407f-b324-c5f0ec7917b2" path="/var/lib/kubelet/pods/e112b6bc-29f1-407f-b324-c5f0ec7917b2/volumes" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.365817 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-fgw98"] Mar 13 10:08:06 crc kubenswrapper[4930]: E0313 10:08:06.366517 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d41547f-e823-475f-8495-ebda890df4a3" containerName="oc" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.366533 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d41547f-e823-475f-8495-ebda890df4a3" containerName="oc" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.366797 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d41547f-e823-475f-8495-ebda890df4a3" containerName="oc" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.369135 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.382141 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fgw98"] Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.489066 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x7sg4\" (UniqueName: \"kubernetes.io/projected/623afded-82a8-46d0-ade8-223b42475163-kube-api-access-x7sg4\") pod \"redhat-marketplace-fgw98\" (UID: \"623afded-82a8-46d0-ade8-223b42475163\") " pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.489229 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/623afded-82a8-46d0-ade8-223b42475163-catalog-content\") pod \"redhat-marketplace-fgw98\" (UID: \"623afded-82a8-46d0-ade8-223b42475163\") " pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.489259 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/623afded-82a8-46d0-ade8-223b42475163-utilities\") pod \"redhat-marketplace-fgw98\" (UID: \"623afded-82a8-46d0-ade8-223b42475163\") " pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.591518 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/623afded-82a8-46d0-ade8-223b42475163-utilities\") pod \"redhat-marketplace-fgw98\" (UID: \"623afded-82a8-46d0-ade8-223b42475163\") " pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.591819 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x7sg4\" (UniqueName: \"kubernetes.io/projected/623afded-82a8-46d0-ade8-223b42475163-kube-api-access-x7sg4\") pod \"redhat-marketplace-fgw98\" (UID: \"623afded-82a8-46d0-ade8-223b42475163\") " pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.591974 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/623afded-82a8-46d0-ade8-223b42475163-catalog-content\") pod \"redhat-marketplace-fgw98\" (UID: \"623afded-82a8-46d0-ade8-223b42475163\") " pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.592046 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/623afded-82a8-46d0-ade8-223b42475163-utilities\") pod \"redhat-marketplace-fgw98\" (UID: \"623afded-82a8-46d0-ade8-223b42475163\") " pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.592516 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/623afded-82a8-46d0-ade8-223b42475163-catalog-content\") pod \"redhat-marketplace-fgw98\" (UID: \"623afded-82a8-46d0-ade8-223b42475163\") " pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.612131 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x7sg4\" (UniqueName: \"kubernetes.io/projected/623afded-82a8-46d0-ade8-223b42475163-kube-api-access-x7sg4\") pod \"redhat-marketplace-fgw98\" (UID: \"623afded-82a8-46d0-ade8-223b42475163\") " pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:06 crc kubenswrapper[4930]: I0313 10:08:06.708423 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:07 crc kubenswrapper[4930]: I0313 10:08:07.998806 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-fgw98"] Mar 13 10:08:08 crc kubenswrapper[4930]: I0313 10:08:08.571618 4930 generic.go:334] "Generic (PLEG): container finished" podID="623afded-82a8-46d0-ade8-223b42475163" containerID="24bbaee43d587db40241e8c06f12889a37e434ab86604031f8bea651ada2dbd8" exitCode=0 Mar 13 10:08:08 crc kubenswrapper[4930]: I0313 10:08:08.572151 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgw98" event={"ID":"623afded-82a8-46d0-ade8-223b42475163","Type":"ContainerDied","Data":"24bbaee43d587db40241e8c06f12889a37e434ab86604031f8bea651ada2dbd8"} Mar 13 10:08:08 crc kubenswrapper[4930]: I0313 10:08:08.572206 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgw98" event={"ID":"623afded-82a8-46d0-ade8-223b42475163","Type":"ContainerStarted","Data":"cf7670c8f8f99819bd37fe242614a840419ee8dae3e4b502fa7ee0ac65c6d774"} Mar 13 10:08:09 crc kubenswrapper[4930]: I0313 10:08:09.585326 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgw98" event={"ID":"623afded-82a8-46d0-ade8-223b42475163","Type":"ContainerStarted","Data":"ba7e933f1aea6ffc1621d83211caec3901499f04f63052b02a096524c8298f7c"} Mar 13 10:08:10 crc kubenswrapper[4930]: I0313 10:08:10.599134 4930 generic.go:334] "Generic (PLEG): container finished" podID="623afded-82a8-46d0-ade8-223b42475163" containerID="ba7e933f1aea6ffc1621d83211caec3901499f04f63052b02a096524c8298f7c" exitCode=0 Mar 13 10:08:10 crc kubenswrapper[4930]: I0313 10:08:10.599248 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgw98" event={"ID":"623afded-82a8-46d0-ade8-223b42475163","Type":"ContainerDied","Data":"ba7e933f1aea6ffc1621d83211caec3901499f04f63052b02a096524c8298f7c"} Mar 13 10:08:11 crc kubenswrapper[4930]: I0313 10:08:11.611501 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgw98" event={"ID":"623afded-82a8-46d0-ade8-223b42475163","Type":"ContainerStarted","Data":"ab226637502d193028d046b3b6b2bb2d9d9f8666ea75bfa69613969d666b78af"} Mar 13 10:08:11 crc kubenswrapper[4930]: I0313 10:08:11.634362 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-fgw98" podStartSLOduration=3.1767449389999998 podStartE2EDuration="5.634338885s" podCreationTimestamp="2026-03-13 10:08:06 +0000 UTC" firstStartedPulling="2026-03-13 10:08:08.574758897 +0000 UTC m=+3329.324673574" lastFinishedPulling="2026-03-13 10:08:11.032352833 +0000 UTC m=+3331.782267520" observedRunningTime="2026-03-13 10:08:11.632703414 +0000 UTC m=+3332.382618081" watchObservedRunningTime="2026-03-13 10:08:11.634338885 +0000 UTC m=+3332.384253572" Mar 13 10:08:12 crc kubenswrapper[4930]: I0313 10:08:12.967683 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:08:13 crc kubenswrapper[4930]: I0313 10:08:13.737714 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xts2v"] Mar 13 10:08:13 crc kubenswrapper[4930]: I0313 10:08:13.738207 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-xts2v" podUID="726723ef-06d8-4873-8d16-383c15165e23" containerName="registry-server" containerID="cri-o://e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362" gracePeriod=2 Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.312051 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.429081 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh5wn\" (UniqueName: \"kubernetes.io/projected/726723ef-06d8-4873-8d16-383c15165e23-kube-api-access-sh5wn\") pod \"726723ef-06d8-4873-8d16-383c15165e23\" (UID: \"726723ef-06d8-4873-8d16-383c15165e23\") " Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.429233 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/726723ef-06d8-4873-8d16-383c15165e23-utilities\") pod \"726723ef-06d8-4873-8d16-383c15165e23\" (UID: \"726723ef-06d8-4873-8d16-383c15165e23\") " Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.429348 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/726723ef-06d8-4873-8d16-383c15165e23-catalog-content\") pod \"726723ef-06d8-4873-8d16-383c15165e23\" (UID: \"726723ef-06d8-4873-8d16-383c15165e23\") " Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.430416 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/726723ef-06d8-4873-8d16-383c15165e23-utilities" (OuterVolumeSpecName: "utilities") pod "726723ef-06d8-4873-8d16-383c15165e23" (UID: "726723ef-06d8-4873-8d16-383c15165e23"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.457603 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/726723ef-06d8-4873-8d16-383c15165e23-kube-api-access-sh5wn" (OuterVolumeSpecName: "kube-api-access-sh5wn") pod "726723ef-06d8-4873-8d16-383c15165e23" (UID: "726723ef-06d8-4873-8d16-383c15165e23"). InnerVolumeSpecName "kube-api-access-sh5wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.488536 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/726723ef-06d8-4873-8d16-383c15165e23-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "726723ef-06d8-4873-8d16-383c15165e23" (UID: "726723ef-06d8-4873-8d16-383c15165e23"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.532155 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh5wn\" (UniqueName: \"kubernetes.io/projected/726723ef-06d8-4873-8d16-383c15165e23-kube-api-access-sh5wn\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.532359 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/726723ef-06d8-4873-8d16-383c15165e23-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.532472 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/726723ef-06d8-4873-8d16-383c15165e23-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.643282 4930 generic.go:334] "Generic (PLEG): container finished" podID="726723ef-06d8-4873-8d16-383c15165e23" containerID="e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362" exitCode=0 Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.643347 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xts2v" event={"ID":"726723ef-06d8-4873-8d16-383c15165e23","Type":"ContainerDied","Data":"e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362"} Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.643384 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-xts2v" event={"ID":"726723ef-06d8-4873-8d16-383c15165e23","Type":"ContainerDied","Data":"aa382d292a3f0985e87ecc2f15ca35eb789c95597730915639f01775fde8f7a3"} Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.643407 4930 scope.go:117] "RemoveContainer" containerID="e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.643801 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-xts2v" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.688693 4930 scope.go:117] "RemoveContainer" containerID="3b6b7481d91fcf99ac7fe598561b0da4930c4763dc54bb3b29b4aed0be75511e" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.694005 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-xts2v"] Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.709963 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-xts2v"] Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.717356 4930 scope.go:117] "RemoveContainer" containerID="f5d2b506009650f27b291cdca51785954b17925291d6e1f2285a147a18b0ef90" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.779760 4930 scope.go:117] "RemoveContainer" containerID="e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362" Mar 13 10:08:14 crc kubenswrapper[4930]: E0313 10:08:14.780265 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362\": container with ID starting with e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362 not found: ID does not exist" containerID="e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.780310 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362"} err="failed to get container status \"e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362\": rpc error: code = NotFound desc = could not find container \"e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362\": container with ID starting with e9ae94e918cb1e9fb5b9d7c43a59960091fd52994ff68447c921f3bebcae7362 not found: ID does not exist" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.780337 4930 scope.go:117] "RemoveContainer" containerID="3b6b7481d91fcf99ac7fe598561b0da4930c4763dc54bb3b29b4aed0be75511e" Mar 13 10:08:14 crc kubenswrapper[4930]: E0313 10:08:14.780752 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b6b7481d91fcf99ac7fe598561b0da4930c4763dc54bb3b29b4aed0be75511e\": container with ID starting with 3b6b7481d91fcf99ac7fe598561b0da4930c4763dc54bb3b29b4aed0be75511e not found: ID does not exist" containerID="3b6b7481d91fcf99ac7fe598561b0da4930c4763dc54bb3b29b4aed0be75511e" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.780771 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b6b7481d91fcf99ac7fe598561b0da4930c4763dc54bb3b29b4aed0be75511e"} err="failed to get container status \"3b6b7481d91fcf99ac7fe598561b0da4930c4763dc54bb3b29b4aed0be75511e\": rpc error: code = NotFound desc = could not find container \"3b6b7481d91fcf99ac7fe598561b0da4930c4763dc54bb3b29b4aed0be75511e\": container with ID starting with 3b6b7481d91fcf99ac7fe598561b0da4930c4763dc54bb3b29b4aed0be75511e not found: ID does not exist" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.780785 4930 scope.go:117] "RemoveContainer" containerID="f5d2b506009650f27b291cdca51785954b17925291d6e1f2285a147a18b0ef90" Mar 13 10:08:14 crc kubenswrapper[4930]: E0313 10:08:14.781024 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f5d2b506009650f27b291cdca51785954b17925291d6e1f2285a147a18b0ef90\": container with ID starting with f5d2b506009650f27b291cdca51785954b17925291d6e1f2285a147a18b0ef90 not found: ID does not exist" containerID="f5d2b506009650f27b291cdca51785954b17925291d6e1f2285a147a18b0ef90" Mar 13 10:08:14 crc kubenswrapper[4930]: I0313 10:08:14.781044 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f5d2b506009650f27b291cdca51785954b17925291d6e1f2285a147a18b0ef90"} err="failed to get container status \"f5d2b506009650f27b291cdca51785954b17925291d6e1f2285a147a18b0ef90\": rpc error: code = NotFound desc = could not find container \"f5d2b506009650f27b291cdca51785954b17925291d6e1f2285a147a18b0ef90\": container with ID starting with f5d2b506009650f27b291cdca51785954b17925291d6e1f2285a147a18b0ef90 not found: ID does not exist" Mar 13 10:08:15 crc kubenswrapper[4930]: I0313 10:08:15.989536 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="726723ef-06d8-4873-8d16-383c15165e23" path="/var/lib/kubelet/pods/726723ef-06d8-4873-8d16-383c15165e23/volumes" Mar 13 10:08:16 crc kubenswrapper[4930]: I0313 10:08:16.711385 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:16 crc kubenswrapper[4930]: I0313 10:08:16.712018 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:16 crc kubenswrapper[4930]: I0313 10:08:16.774327 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:17 crc kubenswrapper[4930]: I0313 10:08:17.732120 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:17 crc kubenswrapper[4930]: I0313 10:08:17.937995 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fgw98"] Mar 13 10:08:19 crc kubenswrapper[4930]: I0313 10:08:19.696387 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-fgw98" podUID="623afded-82a8-46d0-ade8-223b42475163" containerName="registry-server" containerID="cri-o://ab226637502d193028d046b3b6b2bb2d9d9f8666ea75bfa69613969d666b78af" gracePeriod=2 Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.350918 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nk6g7"] Mar 13 10:08:20 crc kubenswrapper[4930]: E0313 10:08:20.351840 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="726723ef-06d8-4873-8d16-383c15165e23" containerName="extract-utilities" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.351864 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="726723ef-06d8-4873-8d16-383c15165e23" containerName="extract-utilities" Mar 13 10:08:20 crc kubenswrapper[4930]: E0313 10:08:20.351884 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="726723ef-06d8-4873-8d16-383c15165e23" containerName="registry-server" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.351893 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="726723ef-06d8-4873-8d16-383c15165e23" containerName="registry-server" Mar 13 10:08:20 crc kubenswrapper[4930]: E0313 10:08:20.351947 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="726723ef-06d8-4873-8d16-383c15165e23" containerName="extract-content" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.351956 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="726723ef-06d8-4873-8d16-383c15165e23" containerName="extract-content" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.352163 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="726723ef-06d8-4873-8d16-383c15165e23" containerName="registry-server" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.353961 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.364090 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nk6g7"] Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.401841 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqmxx\" (UniqueName: \"kubernetes.io/projected/7f276379-33fb-4491-a12f-abb6ce53828d-kube-api-access-nqmxx\") pod \"certified-operators-nk6g7\" (UID: \"7f276379-33fb-4491-a12f-abb6ce53828d\") " pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.401953 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f276379-33fb-4491-a12f-abb6ce53828d-utilities\") pod \"certified-operators-nk6g7\" (UID: \"7f276379-33fb-4491-a12f-abb6ce53828d\") " pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.402025 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f276379-33fb-4491-a12f-abb6ce53828d-catalog-content\") pod \"certified-operators-nk6g7\" (UID: \"7f276379-33fb-4491-a12f-abb6ce53828d\") " pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.504642 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqmxx\" (UniqueName: \"kubernetes.io/projected/7f276379-33fb-4491-a12f-abb6ce53828d-kube-api-access-nqmxx\") pod \"certified-operators-nk6g7\" (UID: \"7f276379-33fb-4491-a12f-abb6ce53828d\") " pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.504795 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f276379-33fb-4491-a12f-abb6ce53828d-utilities\") pod \"certified-operators-nk6g7\" (UID: \"7f276379-33fb-4491-a12f-abb6ce53828d\") " pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.504864 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f276379-33fb-4491-a12f-abb6ce53828d-catalog-content\") pod \"certified-operators-nk6g7\" (UID: \"7f276379-33fb-4491-a12f-abb6ce53828d\") " pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.505549 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f276379-33fb-4491-a12f-abb6ce53828d-utilities\") pod \"certified-operators-nk6g7\" (UID: \"7f276379-33fb-4491-a12f-abb6ce53828d\") " pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.505626 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f276379-33fb-4491-a12f-abb6ce53828d-catalog-content\") pod \"certified-operators-nk6g7\" (UID: \"7f276379-33fb-4491-a12f-abb6ce53828d\") " pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.550509 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqmxx\" (UniqueName: \"kubernetes.io/projected/7f276379-33fb-4491-a12f-abb6ce53828d-kube-api-access-nqmxx\") pod \"certified-operators-nk6g7\" (UID: \"7f276379-33fb-4491-a12f-abb6ce53828d\") " pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.712672 4930 generic.go:334] "Generic (PLEG): container finished" podID="623afded-82a8-46d0-ade8-223b42475163" containerID="ab226637502d193028d046b3b6b2bb2d9d9f8666ea75bfa69613969d666b78af" exitCode=0 Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.712718 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgw98" event={"ID":"623afded-82a8-46d0-ade8-223b42475163","Type":"ContainerDied","Data":"ab226637502d193028d046b3b6b2bb2d9d9f8666ea75bfa69613969d666b78af"} Mar 13 10:08:20 crc kubenswrapper[4930]: I0313 10:08:20.725305 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.007702 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.120700 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/623afded-82a8-46d0-ade8-223b42475163-utilities\") pod \"623afded-82a8-46d0-ade8-223b42475163\" (UID: \"623afded-82a8-46d0-ade8-223b42475163\") " Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.121189 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/623afded-82a8-46d0-ade8-223b42475163-catalog-content\") pod \"623afded-82a8-46d0-ade8-223b42475163\" (UID: \"623afded-82a8-46d0-ade8-223b42475163\") " Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.121338 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7sg4\" (UniqueName: \"kubernetes.io/projected/623afded-82a8-46d0-ade8-223b42475163-kube-api-access-x7sg4\") pod \"623afded-82a8-46d0-ade8-223b42475163\" (UID: \"623afded-82a8-46d0-ade8-223b42475163\") " Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.123883 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/623afded-82a8-46d0-ade8-223b42475163-utilities" (OuterVolumeSpecName: "utilities") pod "623afded-82a8-46d0-ade8-223b42475163" (UID: "623afded-82a8-46d0-ade8-223b42475163"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.132373 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/623afded-82a8-46d0-ade8-223b42475163-kube-api-access-x7sg4" (OuterVolumeSpecName: "kube-api-access-x7sg4") pod "623afded-82a8-46d0-ade8-223b42475163" (UID: "623afded-82a8-46d0-ade8-223b42475163"). InnerVolumeSpecName "kube-api-access-x7sg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.176748 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/623afded-82a8-46d0-ade8-223b42475163-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "623afded-82a8-46d0-ade8-223b42475163" (UID: "623afded-82a8-46d0-ade8-223b42475163"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.226265 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/623afded-82a8-46d0-ade8-223b42475163-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.226300 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/623afded-82a8-46d0-ade8-223b42475163-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.226313 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7sg4\" (UniqueName: \"kubernetes.io/projected/623afded-82a8-46d0-ade8-223b42475163-kube-api-access-x7sg4\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.384076 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nk6g7"] Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.728129 4930 generic.go:334] "Generic (PLEG): container finished" podID="7f276379-33fb-4491-a12f-abb6ce53828d" containerID="756f64b61039ee27e2fa6d4d04fef319e879bd455a584c2e81b3bd77fbd892bb" exitCode=0 Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.728704 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6g7" event={"ID":"7f276379-33fb-4491-a12f-abb6ce53828d","Type":"ContainerDied","Data":"756f64b61039ee27e2fa6d4d04fef319e879bd455a584c2e81b3bd77fbd892bb"} Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.728772 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6g7" event={"ID":"7f276379-33fb-4491-a12f-abb6ce53828d","Type":"ContainerStarted","Data":"80ef05a42a1d350f99c3815564900bacdb3a47b3798663261fcaebd1d4fb4b33"} Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.731825 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-fgw98" event={"ID":"623afded-82a8-46d0-ade8-223b42475163","Type":"ContainerDied","Data":"cf7670c8f8f99819bd37fe242614a840419ee8dae3e4b502fa7ee0ac65c6d774"} Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.731864 4930 scope.go:117] "RemoveContainer" containerID="ab226637502d193028d046b3b6b2bb2d9d9f8666ea75bfa69613969d666b78af" Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.731913 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-fgw98" Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.766686 4930 scope.go:117] "RemoveContainer" containerID="ba7e933f1aea6ffc1621d83211caec3901499f04f63052b02a096524c8298f7c" Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.780949 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-fgw98"] Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.793108 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-fgw98"] Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.806311 4930 scope.go:117] "RemoveContainer" containerID="24bbaee43d587db40241e8c06f12889a37e434ab86604031f8bea651ada2dbd8" Mar 13 10:08:21 crc kubenswrapper[4930]: I0313 10:08:21.988020 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="623afded-82a8-46d0-ade8-223b42475163" path="/var/lib/kubelet/pods/623afded-82a8-46d0-ade8-223b42475163/volumes" Mar 13 10:08:23 crc kubenswrapper[4930]: I0313 10:08:23.769475 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6g7" event={"ID":"7f276379-33fb-4491-a12f-abb6ce53828d","Type":"ContainerStarted","Data":"9b043b0fc3ec4dd4cf85db0d228daa78ecd9e0da3d7a523d329d7015887cfa7b"} Mar 13 10:08:25 crc kubenswrapper[4930]: I0313 10:08:25.793923 4930 generic.go:334] "Generic (PLEG): container finished" podID="7f276379-33fb-4491-a12f-abb6ce53828d" containerID="9b043b0fc3ec4dd4cf85db0d228daa78ecd9e0da3d7a523d329d7015887cfa7b" exitCode=0 Mar 13 10:08:25 crc kubenswrapper[4930]: I0313 10:08:25.794003 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6g7" event={"ID":"7f276379-33fb-4491-a12f-abb6ce53828d","Type":"ContainerDied","Data":"9b043b0fc3ec4dd4cf85db0d228daa78ecd9e0da3d7a523d329d7015887cfa7b"} Mar 13 10:08:26 crc kubenswrapper[4930]: I0313 10:08:26.811615 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6g7" event={"ID":"7f276379-33fb-4491-a12f-abb6ce53828d","Type":"ContainerStarted","Data":"230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603"} Mar 13 10:08:26 crc kubenswrapper[4930]: I0313 10:08:26.844652 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nk6g7" podStartSLOduration=2.330704354 podStartE2EDuration="6.844626718s" podCreationTimestamp="2026-03-13 10:08:20 +0000 UTC" firstStartedPulling="2026-03-13 10:08:21.729886361 +0000 UTC m=+3342.479801038" lastFinishedPulling="2026-03-13 10:08:26.243808725 +0000 UTC m=+3346.993723402" observedRunningTime="2026-03-13 10:08:26.834782361 +0000 UTC m=+3347.584697038" watchObservedRunningTime="2026-03-13 10:08:26.844626718 +0000 UTC m=+3347.594541395" Mar 13 10:08:30 crc kubenswrapper[4930]: I0313 10:08:30.726513 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:30 crc kubenswrapper[4930]: I0313 10:08:30.726971 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:30 crc kubenswrapper[4930]: I0313 10:08:30.781290 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:32 crc kubenswrapper[4930]: I0313 10:08:32.157214 4930 scope.go:117] "RemoveContainer" containerID="007c0f5c70f64210b78ba9fc571927075f2e43025cc25f9e177d9cc437054cf5" Mar 13 10:08:40 crc kubenswrapper[4930]: I0313 10:08:40.784665 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:40 crc kubenswrapper[4930]: I0313 10:08:40.860730 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nk6g7"] Mar 13 10:08:40 crc kubenswrapper[4930]: I0313 10:08:40.955204 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nk6g7" podUID="7f276379-33fb-4491-a12f-abb6ce53828d" containerName="registry-server" containerID="cri-o://230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603" gracePeriod=2 Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.501422 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.682858 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f276379-33fb-4491-a12f-abb6ce53828d-utilities\") pod \"7f276379-33fb-4491-a12f-abb6ce53828d\" (UID: \"7f276379-33fb-4491-a12f-abb6ce53828d\") " Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.683188 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqmxx\" (UniqueName: \"kubernetes.io/projected/7f276379-33fb-4491-a12f-abb6ce53828d-kube-api-access-nqmxx\") pod \"7f276379-33fb-4491-a12f-abb6ce53828d\" (UID: \"7f276379-33fb-4491-a12f-abb6ce53828d\") " Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.683226 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f276379-33fb-4491-a12f-abb6ce53828d-catalog-content\") pod \"7f276379-33fb-4491-a12f-abb6ce53828d\" (UID: \"7f276379-33fb-4491-a12f-abb6ce53828d\") " Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.683920 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f276379-33fb-4491-a12f-abb6ce53828d-utilities" (OuterVolumeSpecName: "utilities") pod "7f276379-33fb-4491-a12f-abb6ce53828d" (UID: "7f276379-33fb-4491-a12f-abb6ce53828d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.691783 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f276379-33fb-4491-a12f-abb6ce53828d-kube-api-access-nqmxx" (OuterVolumeSpecName: "kube-api-access-nqmxx") pod "7f276379-33fb-4491-a12f-abb6ce53828d" (UID: "7f276379-33fb-4491-a12f-abb6ce53828d"). InnerVolumeSpecName "kube-api-access-nqmxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.742461 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7f276379-33fb-4491-a12f-abb6ce53828d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7f276379-33fb-4491-a12f-abb6ce53828d" (UID: "7f276379-33fb-4491-a12f-abb6ce53828d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.786079 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqmxx\" (UniqueName: \"kubernetes.io/projected/7f276379-33fb-4491-a12f-abb6ce53828d-kube-api-access-nqmxx\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.786112 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7f276379-33fb-4491-a12f-abb6ce53828d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.786122 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7f276379-33fb-4491-a12f-abb6ce53828d-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.969102 4930 generic.go:334] "Generic (PLEG): container finished" podID="7f276379-33fb-4491-a12f-abb6ce53828d" containerID="230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603" exitCode=0 Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.969153 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6g7" event={"ID":"7f276379-33fb-4491-a12f-abb6ce53828d","Type":"ContainerDied","Data":"230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603"} Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.969186 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nk6g7" event={"ID":"7f276379-33fb-4491-a12f-abb6ce53828d","Type":"ContainerDied","Data":"80ef05a42a1d350f99c3815564900bacdb3a47b3798663261fcaebd1d4fb4b33"} Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.969206 4930 scope.go:117] "RemoveContainer" containerID="230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603" Mar 13 10:08:41 crc kubenswrapper[4930]: I0313 10:08:41.969605 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nk6g7" Mar 13 10:08:42 crc kubenswrapper[4930]: I0313 10:08:42.011033 4930 scope.go:117] "RemoveContainer" containerID="9b043b0fc3ec4dd4cf85db0d228daa78ecd9e0da3d7a523d329d7015887cfa7b" Mar 13 10:08:42 crc kubenswrapper[4930]: I0313 10:08:42.021556 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nk6g7"] Mar 13 10:08:42 crc kubenswrapper[4930]: I0313 10:08:42.034591 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nk6g7"] Mar 13 10:08:42 crc kubenswrapper[4930]: I0313 10:08:42.049224 4930 scope.go:117] "RemoveContainer" containerID="756f64b61039ee27e2fa6d4d04fef319e879bd455a584c2e81b3bd77fbd892bb" Mar 13 10:08:42 crc kubenswrapper[4930]: I0313 10:08:42.110017 4930 scope.go:117] "RemoveContainer" containerID="230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603" Mar 13 10:08:42 crc kubenswrapper[4930]: E0313 10:08:42.110368 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603\": container with ID starting with 230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603 not found: ID does not exist" containerID="230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603" Mar 13 10:08:42 crc kubenswrapper[4930]: I0313 10:08:42.110406 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603"} err="failed to get container status \"230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603\": rpc error: code = NotFound desc = could not find container \"230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603\": container with ID starting with 230a7b1b05ea02e751ab18342beedca297801b49ee721ac2f17b3f3e8eaa5603 not found: ID does not exist" Mar 13 10:08:42 crc kubenswrapper[4930]: I0313 10:08:42.110454 4930 scope.go:117] "RemoveContainer" containerID="9b043b0fc3ec4dd4cf85db0d228daa78ecd9e0da3d7a523d329d7015887cfa7b" Mar 13 10:08:42 crc kubenswrapper[4930]: E0313 10:08:42.110739 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b043b0fc3ec4dd4cf85db0d228daa78ecd9e0da3d7a523d329d7015887cfa7b\": container with ID starting with 9b043b0fc3ec4dd4cf85db0d228daa78ecd9e0da3d7a523d329d7015887cfa7b not found: ID does not exist" containerID="9b043b0fc3ec4dd4cf85db0d228daa78ecd9e0da3d7a523d329d7015887cfa7b" Mar 13 10:08:42 crc kubenswrapper[4930]: I0313 10:08:42.110771 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b043b0fc3ec4dd4cf85db0d228daa78ecd9e0da3d7a523d329d7015887cfa7b"} err="failed to get container status \"9b043b0fc3ec4dd4cf85db0d228daa78ecd9e0da3d7a523d329d7015887cfa7b\": rpc error: code = NotFound desc = could not find container \"9b043b0fc3ec4dd4cf85db0d228daa78ecd9e0da3d7a523d329d7015887cfa7b\": container with ID starting with 9b043b0fc3ec4dd4cf85db0d228daa78ecd9e0da3d7a523d329d7015887cfa7b not found: ID does not exist" Mar 13 10:08:42 crc kubenswrapper[4930]: I0313 10:08:42.110788 4930 scope.go:117] "RemoveContainer" containerID="756f64b61039ee27e2fa6d4d04fef319e879bd455a584c2e81b3bd77fbd892bb" Mar 13 10:08:42 crc kubenswrapper[4930]: E0313 10:08:42.111061 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"756f64b61039ee27e2fa6d4d04fef319e879bd455a584c2e81b3bd77fbd892bb\": container with ID starting with 756f64b61039ee27e2fa6d4d04fef319e879bd455a584c2e81b3bd77fbd892bb not found: ID does not exist" containerID="756f64b61039ee27e2fa6d4d04fef319e879bd455a584c2e81b3bd77fbd892bb" Mar 13 10:08:42 crc kubenswrapper[4930]: I0313 10:08:42.111121 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"756f64b61039ee27e2fa6d4d04fef319e879bd455a584c2e81b3bd77fbd892bb"} err="failed to get container status \"756f64b61039ee27e2fa6d4d04fef319e879bd455a584c2e81b3bd77fbd892bb\": rpc error: code = NotFound desc = could not find container \"756f64b61039ee27e2fa6d4d04fef319e879bd455a584c2e81b3bd77fbd892bb\": container with ID starting with 756f64b61039ee27e2fa6d4d04fef319e879bd455a584c2e81b3bd77fbd892bb not found: ID does not exist" Mar 13 10:08:43 crc kubenswrapper[4930]: I0313 10:08:43.983483 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f276379-33fb-4491-a12f-abb6ce53828d" path="/var/lib/kubelet/pods/7f276379-33fb-4491-a12f-abb6ce53828d/volumes" Mar 13 10:09:12 crc kubenswrapper[4930]: I0313 10:09:12.308965 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:09:12 crc kubenswrapper[4930]: I0313 10:09:12.309970 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:09:42 crc kubenswrapper[4930]: I0313 10:09:42.308282 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:09:42 crc kubenswrapper[4930]: I0313 10:09:42.308984 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.146782 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556610-6zvcz"] Mar 13 10:10:00 crc kubenswrapper[4930]: E0313 10:10:00.147962 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f276379-33fb-4491-a12f-abb6ce53828d" containerName="registry-server" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.147981 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f276379-33fb-4491-a12f-abb6ce53828d" containerName="registry-server" Mar 13 10:10:00 crc kubenswrapper[4930]: E0313 10:10:00.148011 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="623afded-82a8-46d0-ade8-223b42475163" containerName="extract-content" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.148018 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="623afded-82a8-46d0-ade8-223b42475163" containerName="extract-content" Mar 13 10:10:00 crc kubenswrapper[4930]: E0313 10:10:00.148038 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f276379-33fb-4491-a12f-abb6ce53828d" containerName="extract-content" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.148046 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f276379-33fb-4491-a12f-abb6ce53828d" containerName="extract-content" Mar 13 10:10:00 crc kubenswrapper[4930]: E0313 10:10:00.148062 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="623afded-82a8-46d0-ade8-223b42475163" containerName="registry-server" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.148069 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="623afded-82a8-46d0-ade8-223b42475163" containerName="registry-server" Mar 13 10:10:00 crc kubenswrapper[4930]: E0313 10:10:00.148083 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="623afded-82a8-46d0-ade8-223b42475163" containerName="extract-utilities" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.148090 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="623afded-82a8-46d0-ade8-223b42475163" containerName="extract-utilities" Mar 13 10:10:00 crc kubenswrapper[4930]: E0313 10:10:00.148135 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f276379-33fb-4491-a12f-abb6ce53828d" containerName="extract-utilities" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.148143 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f276379-33fb-4491-a12f-abb6ce53828d" containerName="extract-utilities" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.148410 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="623afded-82a8-46d0-ade8-223b42475163" containerName="registry-server" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.148462 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f276379-33fb-4491-a12f-abb6ce53828d" containerName="registry-server" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.149496 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556610-6zvcz" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.151861 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.152392 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.153068 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.160358 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556610-6zvcz"] Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.325102 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5cwt\" (UniqueName: \"kubernetes.io/projected/2c020e7e-0ae4-4c30-b625-a2afc1ec5891-kube-api-access-w5cwt\") pod \"auto-csr-approver-29556610-6zvcz\" (UID: \"2c020e7e-0ae4-4c30-b625-a2afc1ec5891\") " pod="openshift-infra/auto-csr-approver-29556610-6zvcz" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.427549 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5cwt\" (UniqueName: \"kubernetes.io/projected/2c020e7e-0ae4-4c30-b625-a2afc1ec5891-kube-api-access-w5cwt\") pod \"auto-csr-approver-29556610-6zvcz\" (UID: \"2c020e7e-0ae4-4c30-b625-a2afc1ec5891\") " pod="openshift-infra/auto-csr-approver-29556610-6zvcz" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.445975 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5cwt\" (UniqueName: \"kubernetes.io/projected/2c020e7e-0ae4-4c30-b625-a2afc1ec5891-kube-api-access-w5cwt\") pod \"auto-csr-approver-29556610-6zvcz\" (UID: \"2c020e7e-0ae4-4c30-b625-a2afc1ec5891\") " pod="openshift-infra/auto-csr-approver-29556610-6zvcz" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.472964 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556610-6zvcz" Mar 13 10:10:00 crc kubenswrapper[4930]: I0313 10:10:00.950363 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556610-6zvcz"] Mar 13 10:10:01 crc kubenswrapper[4930]: I0313 10:10:01.221457 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556610-6zvcz" event={"ID":"2c020e7e-0ae4-4c30-b625-a2afc1ec5891","Type":"ContainerStarted","Data":"4a6f984b73e697c69ec3149e8fefbe6f8ac065a837d48a210d0485bc5ac2799d"} Mar 13 10:10:03 crc kubenswrapper[4930]: I0313 10:10:03.248298 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556610-6zvcz" event={"ID":"2c020e7e-0ae4-4c30-b625-a2afc1ec5891","Type":"ContainerStarted","Data":"7edfd81c58a94c68ba605930c9cae1138728751b1f581f0228adff0db7b05a6d"} Mar 13 10:10:03 crc kubenswrapper[4930]: I0313 10:10:03.262794 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556610-6zvcz" podStartSLOduration=1.595007448 podStartE2EDuration="3.262776984s" podCreationTimestamp="2026-03-13 10:10:00 +0000 UTC" firstStartedPulling="2026-03-13 10:10:00.952699829 +0000 UTC m=+3441.702614496" lastFinishedPulling="2026-03-13 10:10:02.620469355 +0000 UTC m=+3443.370384032" observedRunningTime="2026-03-13 10:10:03.261318738 +0000 UTC m=+3444.011233415" watchObservedRunningTime="2026-03-13 10:10:03.262776984 +0000 UTC m=+3444.012691661" Mar 13 10:10:04 crc kubenswrapper[4930]: I0313 10:10:04.260636 4930 generic.go:334] "Generic (PLEG): container finished" podID="2c020e7e-0ae4-4c30-b625-a2afc1ec5891" containerID="7edfd81c58a94c68ba605930c9cae1138728751b1f581f0228adff0db7b05a6d" exitCode=0 Mar 13 10:10:04 crc kubenswrapper[4930]: I0313 10:10:04.260688 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556610-6zvcz" event={"ID":"2c020e7e-0ae4-4c30-b625-a2afc1ec5891","Type":"ContainerDied","Data":"7edfd81c58a94c68ba605930c9cae1138728751b1f581f0228adff0db7b05a6d"} Mar 13 10:10:05 crc kubenswrapper[4930]: I0313 10:10:05.737934 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556610-6zvcz" Mar 13 10:10:05 crc kubenswrapper[4930]: I0313 10:10:05.766803 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5cwt\" (UniqueName: \"kubernetes.io/projected/2c020e7e-0ae4-4c30-b625-a2afc1ec5891-kube-api-access-w5cwt\") pod \"2c020e7e-0ae4-4c30-b625-a2afc1ec5891\" (UID: \"2c020e7e-0ae4-4c30-b625-a2afc1ec5891\") " Mar 13 10:10:05 crc kubenswrapper[4930]: I0313 10:10:05.776083 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c020e7e-0ae4-4c30-b625-a2afc1ec5891-kube-api-access-w5cwt" (OuterVolumeSpecName: "kube-api-access-w5cwt") pod "2c020e7e-0ae4-4c30-b625-a2afc1ec5891" (UID: "2c020e7e-0ae4-4c30-b625-a2afc1ec5891"). InnerVolumeSpecName "kube-api-access-w5cwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:10:05 crc kubenswrapper[4930]: I0313 10:10:05.872704 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5cwt\" (UniqueName: \"kubernetes.io/projected/2c020e7e-0ae4-4c30-b625-a2afc1ec5891-kube-api-access-w5cwt\") on node \"crc\" DevicePath \"\"" Mar 13 10:10:06 crc kubenswrapper[4930]: I0313 10:10:06.283360 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556610-6zvcz" event={"ID":"2c020e7e-0ae4-4c30-b625-a2afc1ec5891","Type":"ContainerDied","Data":"4a6f984b73e697c69ec3149e8fefbe6f8ac065a837d48a210d0485bc5ac2799d"} Mar 13 10:10:06 crc kubenswrapper[4930]: I0313 10:10:06.283421 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a6f984b73e697c69ec3149e8fefbe6f8ac065a837d48a210d0485bc5ac2799d" Mar 13 10:10:06 crc kubenswrapper[4930]: I0313 10:10:06.283484 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556610-6zvcz" Mar 13 10:10:06 crc kubenswrapper[4930]: I0313 10:10:06.367067 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556604-k5xm4"] Mar 13 10:10:06 crc kubenswrapper[4930]: I0313 10:10:06.387221 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556604-k5xm4"] Mar 13 10:10:07 crc kubenswrapper[4930]: I0313 10:10:07.989497 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca1762ed-5f30-42a1-b549-ad60e97dfa8d" path="/var/lib/kubelet/pods/ca1762ed-5f30-42a1-b549-ad60e97dfa8d/volumes" Mar 13 10:10:12 crc kubenswrapper[4930]: I0313 10:10:12.308937 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:10:12 crc kubenswrapper[4930]: I0313 10:10:12.309885 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:10:12 crc kubenswrapper[4930]: I0313 10:10:12.309955 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 10:10:12 crc kubenswrapper[4930]: I0313 10:10:12.311220 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f728726a2be84d2953c1136412cbb0280b9a681f89803670792f59768fcb7bbe"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:10:12 crc kubenswrapper[4930]: I0313 10:10:12.311290 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://f728726a2be84d2953c1136412cbb0280b9a681f89803670792f59768fcb7bbe" gracePeriod=600 Mar 13 10:10:13 crc kubenswrapper[4930]: I0313 10:10:13.364018 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="f728726a2be84d2953c1136412cbb0280b9a681f89803670792f59768fcb7bbe" exitCode=0 Mar 13 10:10:13 crc kubenswrapper[4930]: I0313 10:10:13.364048 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"f728726a2be84d2953c1136412cbb0280b9a681f89803670792f59768fcb7bbe"} Mar 13 10:10:13 crc kubenswrapper[4930]: I0313 10:10:13.364741 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf"} Mar 13 10:10:13 crc kubenswrapper[4930]: I0313 10:10:13.364760 4930 scope.go:117] "RemoveContainer" containerID="d996c1f6213e08e90afb2fbfd5aebadf05aa98478cdbf6d8fea6088da445fe38" Mar 13 10:10:32 crc kubenswrapper[4930]: I0313 10:10:32.330911 4930 scope.go:117] "RemoveContainer" containerID="6dec1d46b82212ef510faaa384c55d92199cb59edc4a7bae0fb9bf4aba9be3af" Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.146262 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556612-88shc"] Mar 13 10:12:00 crc kubenswrapper[4930]: E0313 10:12:00.147462 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c020e7e-0ae4-4c30-b625-a2afc1ec5891" containerName="oc" Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.147479 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c020e7e-0ae4-4c30-b625-a2afc1ec5891" containerName="oc" Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.147717 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c020e7e-0ae4-4c30-b625-a2afc1ec5891" containerName="oc" Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.148569 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556612-88shc" Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.150332 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.152019 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.152173 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.182775 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556612-88shc"] Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.240888 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6xplv\" (UniqueName: \"kubernetes.io/projected/afdb2c21-1f79-4d0f-a99d-d72b782ccbf5-kube-api-access-6xplv\") pod \"auto-csr-approver-29556612-88shc\" (UID: \"afdb2c21-1f79-4d0f-a99d-d72b782ccbf5\") " pod="openshift-infra/auto-csr-approver-29556612-88shc" Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.343663 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6xplv\" (UniqueName: \"kubernetes.io/projected/afdb2c21-1f79-4d0f-a99d-d72b782ccbf5-kube-api-access-6xplv\") pod \"auto-csr-approver-29556612-88shc\" (UID: \"afdb2c21-1f79-4d0f-a99d-d72b782ccbf5\") " pod="openshift-infra/auto-csr-approver-29556612-88shc" Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.362954 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6xplv\" (UniqueName: \"kubernetes.io/projected/afdb2c21-1f79-4d0f-a99d-d72b782ccbf5-kube-api-access-6xplv\") pod \"auto-csr-approver-29556612-88shc\" (UID: \"afdb2c21-1f79-4d0f-a99d-d72b782ccbf5\") " pod="openshift-infra/auto-csr-approver-29556612-88shc" Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.468258 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556612-88shc" Mar 13 10:12:00 crc kubenswrapper[4930]: I0313 10:12:00.935378 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556612-88shc"] Mar 13 10:12:01 crc kubenswrapper[4930]: I0313 10:12:01.605001 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556612-88shc" event={"ID":"afdb2c21-1f79-4d0f-a99d-d72b782ccbf5","Type":"ContainerStarted","Data":"3c689d2629b511cc87df08ba25c442bab24d48ad32b1296bb8235bf150ef0a41"} Mar 13 10:12:02 crc kubenswrapper[4930]: I0313 10:12:02.615761 4930 generic.go:334] "Generic (PLEG): container finished" podID="afdb2c21-1f79-4d0f-a99d-d72b782ccbf5" containerID="1592e8677b50e62c256a8c8f25273e374912dcbfad982e344e17d3bdd694fe61" exitCode=0 Mar 13 10:12:02 crc kubenswrapper[4930]: I0313 10:12:02.615853 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556612-88shc" event={"ID":"afdb2c21-1f79-4d0f-a99d-d72b782ccbf5","Type":"ContainerDied","Data":"1592e8677b50e62c256a8c8f25273e374912dcbfad982e344e17d3bdd694fe61"} Mar 13 10:12:04 crc kubenswrapper[4930]: I0313 10:12:04.094334 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556612-88shc" Mar 13 10:12:04 crc kubenswrapper[4930]: I0313 10:12:04.170129 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6xplv\" (UniqueName: \"kubernetes.io/projected/afdb2c21-1f79-4d0f-a99d-d72b782ccbf5-kube-api-access-6xplv\") pod \"afdb2c21-1f79-4d0f-a99d-d72b782ccbf5\" (UID: \"afdb2c21-1f79-4d0f-a99d-d72b782ccbf5\") " Mar 13 10:12:04 crc kubenswrapper[4930]: I0313 10:12:04.180515 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afdb2c21-1f79-4d0f-a99d-d72b782ccbf5-kube-api-access-6xplv" (OuterVolumeSpecName: "kube-api-access-6xplv") pod "afdb2c21-1f79-4d0f-a99d-d72b782ccbf5" (UID: "afdb2c21-1f79-4d0f-a99d-d72b782ccbf5"). InnerVolumeSpecName "kube-api-access-6xplv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:12:04 crc kubenswrapper[4930]: I0313 10:12:04.273752 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6xplv\" (UniqueName: \"kubernetes.io/projected/afdb2c21-1f79-4d0f-a99d-d72b782ccbf5-kube-api-access-6xplv\") on node \"crc\" DevicePath \"\"" Mar 13 10:12:04 crc kubenswrapper[4930]: I0313 10:12:04.638164 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556612-88shc" event={"ID":"afdb2c21-1f79-4d0f-a99d-d72b782ccbf5","Type":"ContainerDied","Data":"3c689d2629b511cc87df08ba25c442bab24d48ad32b1296bb8235bf150ef0a41"} Mar 13 10:12:04 crc kubenswrapper[4930]: I0313 10:12:04.638565 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c689d2629b511cc87df08ba25c442bab24d48ad32b1296bb8235bf150ef0a41" Mar 13 10:12:04 crc kubenswrapper[4930]: I0313 10:12:04.638383 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556612-88shc" Mar 13 10:12:05 crc kubenswrapper[4930]: I0313 10:12:05.175369 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556606-klqjt"] Mar 13 10:12:05 crc kubenswrapper[4930]: I0313 10:12:05.189504 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556606-klqjt"] Mar 13 10:12:05 crc kubenswrapper[4930]: I0313 10:12:05.987382 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01c99807-49ba-4ee9-a14a-f746b6f00e62" path="/var/lib/kubelet/pods/01c99807-49ba-4ee9-a14a-f746b6f00e62/volumes" Mar 13 10:12:12 crc kubenswrapper[4930]: I0313 10:12:12.308205 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:12:12 crc kubenswrapper[4930]: I0313 10:12:12.308843 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:12:32 crc kubenswrapper[4930]: I0313 10:12:32.432496 4930 scope.go:117] "RemoveContainer" containerID="aa281c345ef0a3caa0b8dfc91b5d14b6c4ec6f0d3add501b2738d0cd1495273f" Mar 13 10:12:42 crc kubenswrapper[4930]: I0313 10:12:42.308089 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:12:42 crc kubenswrapper[4930]: I0313 10:12:42.308645 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:13:12 crc kubenswrapper[4930]: I0313 10:13:12.308674 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:13:12 crc kubenswrapper[4930]: I0313 10:13:12.309104 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:13:12 crc kubenswrapper[4930]: I0313 10:13:12.309147 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 10:13:12 crc kubenswrapper[4930]: I0313 10:13:12.309984 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:13:12 crc kubenswrapper[4930]: I0313 10:13:12.310068 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" gracePeriod=600 Mar 13 10:13:12 crc kubenswrapper[4930]: E0313 10:13:12.457649 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:13:13 crc kubenswrapper[4930]: I0313 10:13:13.360158 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" exitCode=0 Mar 13 10:13:13 crc kubenswrapper[4930]: I0313 10:13:13.361417 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf"} Mar 13 10:13:13 crc kubenswrapper[4930]: I0313 10:13:13.361607 4930 scope.go:117] "RemoveContainer" containerID="f728726a2be84d2953c1136412cbb0280b9a681f89803670792f59768fcb7bbe" Mar 13 10:13:13 crc kubenswrapper[4930]: I0313 10:13:13.362667 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:13:13 crc kubenswrapper[4930]: E0313 10:13:13.363284 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:13:27 crc kubenswrapper[4930]: I0313 10:13:27.970778 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:13:27 crc kubenswrapper[4930]: E0313 10:13:27.971715 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:13:39 crc kubenswrapper[4930]: I0313 10:13:39.983847 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:13:39 crc kubenswrapper[4930]: E0313 10:13:39.985273 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:13:53 crc kubenswrapper[4930]: I0313 10:13:53.971226 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:13:53 crc kubenswrapper[4930]: E0313 10:13:53.972166 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.153764 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556614-rc9f4"] Mar 13 10:14:00 crc kubenswrapper[4930]: E0313 10:14:00.155683 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afdb2c21-1f79-4d0f-a99d-d72b782ccbf5" containerName="oc" Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.155704 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="afdb2c21-1f79-4d0f-a99d-d72b782ccbf5" containerName="oc" Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.156003 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="afdb2c21-1f79-4d0f-a99d-d72b782ccbf5" containerName="oc" Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.157161 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556614-rc9f4" Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.159561 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.160033 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.160593 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.170144 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556614-rc9f4"] Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.335492 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vldxr\" (UniqueName: \"kubernetes.io/projected/c4b1588b-1d42-41c4-ab24-35ae3edea14f-kube-api-access-vldxr\") pod \"auto-csr-approver-29556614-rc9f4\" (UID: \"c4b1588b-1d42-41c4-ab24-35ae3edea14f\") " pod="openshift-infra/auto-csr-approver-29556614-rc9f4" Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.438166 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vldxr\" (UniqueName: \"kubernetes.io/projected/c4b1588b-1d42-41c4-ab24-35ae3edea14f-kube-api-access-vldxr\") pod \"auto-csr-approver-29556614-rc9f4\" (UID: \"c4b1588b-1d42-41c4-ab24-35ae3edea14f\") " pod="openshift-infra/auto-csr-approver-29556614-rc9f4" Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.457040 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vldxr\" (UniqueName: \"kubernetes.io/projected/c4b1588b-1d42-41c4-ab24-35ae3edea14f-kube-api-access-vldxr\") pod \"auto-csr-approver-29556614-rc9f4\" (UID: \"c4b1588b-1d42-41c4-ab24-35ae3edea14f\") " pod="openshift-infra/auto-csr-approver-29556614-rc9f4" Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.481034 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556614-rc9f4" Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.954141 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556614-rc9f4"] Mar 13 10:14:00 crc kubenswrapper[4930]: I0313 10:14:00.956329 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:14:01 crc kubenswrapper[4930]: I0313 10:14:01.864738 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556614-rc9f4" event={"ID":"c4b1588b-1d42-41c4-ab24-35ae3edea14f","Type":"ContainerStarted","Data":"6afae338032e6a7784e7190829102f54fb672cd9f8ff769855a8fb4377395a48"} Mar 13 10:14:02 crc kubenswrapper[4930]: I0313 10:14:02.878073 4930 generic.go:334] "Generic (PLEG): container finished" podID="c4b1588b-1d42-41c4-ab24-35ae3edea14f" containerID="8c3b57e581f809ab8fdbf2a8a1c5e433242a002245cbf8aaf34b54b70566cfc4" exitCode=0 Mar 13 10:14:02 crc kubenswrapper[4930]: I0313 10:14:02.878127 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556614-rc9f4" event={"ID":"c4b1588b-1d42-41c4-ab24-35ae3edea14f","Type":"ContainerDied","Data":"8c3b57e581f809ab8fdbf2a8a1c5e433242a002245cbf8aaf34b54b70566cfc4"} Mar 13 10:14:04 crc kubenswrapper[4930]: I0313 10:14:04.794457 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556614-rc9f4" Mar 13 10:14:04 crc kubenswrapper[4930]: I0313 10:14:04.900718 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556614-rc9f4" event={"ID":"c4b1588b-1d42-41c4-ab24-35ae3edea14f","Type":"ContainerDied","Data":"6afae338032e6a7784e7190829102f54fb672cd9f8ff769855a8fb4377395a48"} Mar 13 10:14:04 crc kubenswrapper[4930]: I0313 10:14:04.900764 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6afae338032e6a7784e7190829102f54fb672cd9f8ff769855a8fb4377395a48" Mar 13 10:14:04 crc kubenswrapper[4930]: I0313 10:14:04.900766 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556614-rc9f4" Mar 13 10:14:04 crc kubenswrapper[4930]: I0313 10:14:04.950794 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vldxr\" (UniqueName: \"kubernetes.io/projected/c4b1588b-1d42-41c4-ab24-35ae3edea14f-kube-api-access-vldxr\") pod \"c4b1588b-1d42-41c4-ab24-35ae3edea14f\" (UID: \"c4b1588b-1d42-41c4-ab24-35ae3edea14f\") " Mar 13 10:14:04 crc kubenswrapper[4930]: I0313 10:14:04.967614 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4b1588b-1d42-41c4-ab24-35ae3edea14f-kube-api-access-vldxr" (OuterVolumeSpecName: "kube-api-access-vldxr") pod "c4b1588b-1d42-41c4-ab24-35ae3edea14f" (UID: "c4b1588b-1d42-41c4-ab24-35ae3edea14f"). InnerVolumeSpecName "kube-api-access-vldxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:14:04 crc kubenswrapper[4930]: I0313 10:14:04.972784 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:14:04 crc kubenswrapper[4930]: E0313 10:14:04.973162 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:14:05 crc kubenswrapper[4930]: I0313 10:14:05.055462 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vldxr\" (UniqueName: \"kubernetes.io/projected/c4b1588b-1d42-41c4-ab24-35ae3edea14f-kube-api-access-vldxr\") on node \"crc\" DevicePath \"\"" Mar 13 10:14:05 crc kubenswrapper[4930]: I0313 10:14:05.874562 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556608-zczc8"] Mar 13 10:14:05 crc kubenswrapper[4930]: I0313 10:14:05.887544 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556608-zczc8"] Mar 13 10:14:05 crc kubenswrapper[4930]: I0313 10:14:05.992624 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d41547f-e823-475f-8495-ebda890df4a3" path="/var/lib/kubelet/pods/3d41547f-e823-475f-8495-ebda890df4a3/volumes" Mar 13 10:14:16 crc kubenswrapper[4930]: I0313 10:14:16.971145 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:14:16 crc kubenswrapper[4930]: E0313 10:14:16.972055 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:14:27 crc kubenswrapper[4930]: I0313 10:14:27.975342 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:14:27 crc kubenswrapper[4930]: E0313 10:14:27.976365 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:14:32 crc kubenswrapper[4930]: I0313 10:14:32.530955 4930 scope.go:117] "RemoveContainer" containerID="b070893d9a53bfba06eea2e119897fe6ea9c1054bb9913730f6260e6cb98716d" Mar 13 10:14:40 crc kubenswrapper[4930]: I0313 10:14:40.971448 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:14:40 crc kubenswrapper[4930]: E0313 10:14:40.973737 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:14:54 crc kubenswrapper[4930]: I0313 10:14:54.972025 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:14:54 crc kubenswrapper[4930]: E0313 10:14:54.973021 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.150607 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk"] Mar 13 10:15:00 crc kubenswrapper[4930]: E0313 10:15:00.151517 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4b1588b-1d42-41c4-ab24-35ae3edea14f" containerName="oc" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.151534 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4b1588b-1d42-41c4-ab24-35ae3edea14f" containerName="oc" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.151813 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4b1588b-1d42-41c4-ab24-35ae3edea14f" containerName="oc" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.152704 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.155606 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.155832 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.162274 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk"] Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.301340 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f24fcf27-dc35-4210-9681-5bba045fd47d-secret-volume\") pod \"collect-profiles-29556615-7d4hk\" (UID: \"f24fcf27-dc35-4210-9681-5bba045fd47d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.301468 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s85cb\" (UniqueName: \"kubernetes.io/projected/f24fcf27-dc35-4210-9681-5bba045fd47d-kube-api-access-s85cb\") pod \"collect-profiles-29556615-7d4hk\" (UID: \"f24fcf27-dc35-4210-9681-5bba045fd47d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.301640 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f24fcf27-dc35-4210-9681-5bba045fd47d-config-volume\") pod \"collect-profiles-29556615-7d4hk\" (UID: \"f24fcf27-dc35-4210-9681-5bba045fd47d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.404049 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s85cb\" (UniqueName: \"kubernetes.io/projected/f24fcf27-dc35-4210-9681-5bba045fd47d-kube-api-access-s85cb\") pod \"collect-profiles-29556615-7d4hk\" (UID: \"f24fcf27-dc35-4210-9681-5bba045fd47d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.404846 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f24fcf27-dc35-4210-9681-5bba045fd47d-config-volume\") pod \"collect-profiles-29556615-7d4hk\" (UID: \"f24fcf27-dc35-4210-9681-5bba045fd47d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.405211 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f24fcf27-dc35-4210-9681-5bba045fd47d-secret-volume\") pod \"collect-profiles-29556615-7d4hk\" (UID: \"f24fcf27-dc35-4210-9681-5bba045fd47d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.405742 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f24fcf27-dc35-4210-9681-5bba045fd47d-config-volume\") pod \"collect-profiles-29556615-7d4hk\" (UID: \"f24fcf27-dc35-4210-9681-5bba045fd47d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.411259 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f24fcf27-dc35-4210-9681-5bba045fd47d-secret-volume\") pod \"collect-profiles-29556615-7d4hk\" (UID: \"f24fcf27-dc35-4210-9681-5bba045fd47d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.422201 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s85cb\" (UniqueName: \"kubernetes.io/projected/f24fcf27-dc35-4210-9681-5bba045fd47d-kube-api-access-s85cb\") pod \"collect-profiles-29556615-7d4hk\" (UID: \"f24fcf27-dc35-4210-9681-5bba045fd47d\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.480599 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:00 crc kubenswrapper[4930]: I0313 10:15:00.988918 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk"] Mar 13 10:15:01 crc kubenswrapper[4930]: I0313 10:15:01.513035 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" event={"ID":"f24fcf27-dc35-4210-9681-5bba045fd47d","Type":"ContainerStarted","Data":"9129f162d74e1d86ec716d9bdf60bf980689904e44932a7c70e4bc6a23e17353"} Mar 13 10:15:01 crc kubenswrapper[4930]: I0313 10:15:01.513462 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" event={"ID":"f24fcf27-dc35-4210-9681-5bba045fd47d","Type":"ContainerStarted","Data":"095f6328180222c86f6f6bb2fd494388605c2cec567d857ee8dfab8423a9c571"} Mar 13 10:15:01 crc kubenswrapper[4930]: I0313 10:15:01.536323 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" podStartSLOduration=1.536295843 podStartE2EDuration="1.536295843s" podCreationTimestamp="2026-03-13 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:15:01.52662426 +0000 UTC m=+3742.276538947" watchObservedRunningTime="2026-03-13 10:15:01.536295843 +0000 UTC m=+3742.286210520" Mar 13 10:15:02 crc kubenswrapper[4930]: I0313 10:15:02.524935 4930 generic.go:334] "Generic (PLEG): container finished" podID="f24fcf27-dc35-4210-9681-5bba045fd47d" containerID="9129f162d74e1d86ec716d9bdf60bf980689904e44932a7c70e4bc6a23e17353" exitCode=0 Mar 13 10:15:02 crc kubenswrapper[4930]: I0313 10:15:02.525274 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" event={"ID":"f24fcf27-dc35-4210-9681-5bba045fd47d","Type":"ContainerDied","Data":"9129f162d74e1d86ec716d9bdf60bf980689904e44932a7c70e4bc6a23e17353"} Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.036345 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.133390 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f24fcf27-dc35-4210-9681-5bba045fd47d-config-volume\") pod \"f24fcf27-dc35-4210-9681-5bba045fd47d\" (UID: \"f24fcf27-dc35-4210-9681-5bba045fd47d\") " Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.133597 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s85cb\" (UniqueName: \"kubernetes.io/projected/f24fcf27-dc35-4210-9681-5bba045fd47d-kube-api-access-s85cb\") pod \"f24fcf27-dc35-4210-9681-5bba045fd47d\" (UID: \"f24fcf27-dc35-4210-9681-5bba045fd47d\") " Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.133797 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f24fcf27-dc35-4210-9681-5bba045fd47d-secret-volume\") pod \"f24fcf27-dc35-4210-9681-5bba045fd47d\" (UID: \"f24fcf27-dc35-4210-9681-5bba045fd47d\") " Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.134475 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f24fcf27-dc35-4210-9681-5bba045fd47d-config-volume" (OuterVolumeSpecName: "config-volume") pod "f24fcf27-dc35-4210-9681-5bba045fd47d" (UID: "f24fcf27-dc35-4210-9681-5bba045fd47d"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.140194 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f24fcf27-dc35-4210-9681-5bba045fd47d-kube-api-access-s85cb" (OuterVolumeSpecName: "kube-api-access-s85cb") pod "f24fcf27-dc35-4210-9681-5bba045fd47d" (UID: "f24fcf27-dc35-4210-9681-5bba045fd47d"). InnerVolumeSpecName "kube-api-access-s85cb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.142651 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f24fcf27-dc35-4210-9681-5bba045fd47d-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f24fcf27-dc35-4210-9681-5bba045fd47d" (UID: "f24fcf27-dc35-4210-9681-5bba045fd47d"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.236672 4930 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f24fcf27-dc35-4210-9681-5bba045fd47d-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.236705 4930 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f24fcf27-dc35-4210-9681-5bba045fd47d-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.236718 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s85cb\" (UniqueName: \"kubernetes.io/projected/f24fcf27-dc35-4210-9681-5bba045fd47d-kube-api-access-s85cb\") on node \"crc\" DevicePath \"\"" Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.557765 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" event={"ID":"f24fcf27-dc35-4210-9681-5bba045fd47d","Type":"ContainerDied","Data":"095f6328180222c86f6f6bb2fd494388605c2cec567d857ee8dfab8423a9c571"} Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.557820 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="095f6328180222c86f6f6bb2fd494388605c2cec567d857ee8dfab8423a9c571" Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.557916 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556615-7d4hk" Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.636294 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49"] Mar 13 10:15:04 crc kubenswrapper[4930]: I0313 10:15:04.648590 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556570-gmd49"] Mar 13 10:15:05 crc kubenswrapper[4930]: I0313 10:15:05.984726 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25af5d21-ed16-4f4e-b49c-c78b965bc965" path="/var/lib/kubelet/pods/25af5d21-ed16-4f4e-b49c-c78b965bc965/volumes" Mar 13 10:15:08 crc kubenswrapper[4930]: I0313 10:15:08.971773 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:15:08 crc kubenswrapper[4930]: E0313 10:15:08.973992 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:15:22 crc kubenswrapper[4930]: I0313 10:15:22.972319 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:15:22 crc kubenswrapper[4930]: E0313 10:15:22.973606 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.576217 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bnrs5"] Mar 13 10:15:26 crc kubenswrapper[4930]: E0313 10:15:26.577328 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f24fcf27-dc35-4210-9681-5bba045fd47d" containerName="collect-profiles" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.577342 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f24fcf27-dc35-4210-9681-5bba045fd47d" containerName="collect-profiles" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.577680 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f24fcf27-dc35-4210-9681-5bba045fd47d" containerName="collect-profiles" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.579543 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.595897 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bnrs5"] Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.632662 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2973615f-255e-4e3a-839d-dff33bd7fb1e-utilities\") pod \"redhat-operators-bnrs5\" (UID: \"2973615f-255e-4e3a-839d-dff33bd7fb1e\") " pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.632713 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwh42\" (UniqueName: \"kubernetes.io/projected/2973615f-255e-4e3a-839d-dff33bd7fb1e-kube-api-access-pwh42\") pod \"redhat-operators-bnrs5\" (UID: \"2973615f-255e-4e3a-839d-dff33bd7fb1e\") " pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.632800 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2973615f-255e-4e3a-839d-dff33bd7fb1e-catalog-content\") pod \"redhat-operators-bnrs5\" (UID: \"2973615f-255e-4e3a-839d-dff33bd7fb1e\") " pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.734905 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2973615f-255e-4e3a-839d-dff33bd7fb1e-catalog-content\") pod \"redhat-operators-bnrs5\" (UID: \"2973615f-255e-4e3a-839d-dff33bd7fb1e\") " pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.735068 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2973615f-255e-4e3a-839d-dff33bd7fb1e-utilities\") pod \"redhat-operators-bnrs5\" (UID: \"2973615f-255e-4e3a-839d-dff33bd7fb1e\") " pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.735093 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwh42\" (UniqueName: \"kubernetes.io/projected/2973615f-255e-4e3a-839d-dff33bd7fb1e-kube-api-access-pwh42\") pod \"redhat-operators-bnrs5\" (UID: \"2973615f-255e-4e3a-839d-dff33bd7fb1e\") " pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.735448 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2973615f-255e-4e3a-839d-dff33bd7fb1e-catalog-content\") pod \"redhat-operators-bnrs5\" (UID: \"2973615f-255e-4e3a-839d-dff33bd7fb1e\") " pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.735727 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2973615f-255e-4e3a-839d-dff33bd7fb1e-utilities\") pod \"redhat-operators-bnrs5\" (UID: \"2973615f-255e-4e3a-839d-dff33bd7fb1e\") " pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.753520 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwh42\" (UniqueName: \"kubernetes.io/projected/2973615f-255e-4e3a-839d-dff33bd7fb1e-kube-api-access-pwh42\") pod \"redhat-operators-bnrs5\" (UID: \"2973615f-255e-4e3a-839d-dff33bd7fb1e\") " pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:26 crc kubenswrapper[4930]: I0313 10:15:26.907296 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:27 crc kubenswrapper[4930]: I0313 10:15:27.345399 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bnrs5"] Mar 13 10:15:27 crc kubenswrapper[4930]: I0313 10:15:27.833198 4930 generic.go:334] "Generic (PLEG): container finished" podID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerID="b1aefee2a21791715c6bb68a14631bfb5b031b5d5af9989b61f48d491ea68012" exitCode=0 Mar 13 10:15:27 crc kubenswrapper[4930]: I0313 10:15:27.833245 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnrs5" event={"ID":"2973615f-255e-4e3a-839d-dff33bd7fb1e","Type":"ContainerDied","Data":"b1aefee2a21791715c6bb68a14631bfb5b031b5d5af9989b61f48d491ea68012"} Mar 13 10:15:27 crc kubenswrapper[4930]: I0313 10:15:27.833269 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnrs5" event={"ID":"2973615f-255e-4e3a-839d-dff33bd7fb1e","Type":"ContainerStarted","Data":"e2a1c2eb9dd663d4282c92162e4c31e0d663ff53bb40e7893136efff4e1e8ae5"} Mar 13 10:15:29 crc kubenswrapper[4930]: I0313 10:15:29.856714 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnrs5" event={"ID":"2973615f-255e-4e3a-839d-dff33bd7fb1e","Type":"ContainerStarted","Data":"a1fad07143b6fe243430cacd47d4274b8c0136523ab869e847b3a15de19c19f9"} Mar 13 10:15:32 crc kubenswrapper[4930]: I0313 10:15:32.612013 4930 scope.go:117] "RemoveContainer" containerID="c3f6cdc3f893b1ef270098459978b42705f2b9a5459595812e1a5430879a177b" Mar 13 10:15:34 crc kubenswrapper[4930]: I0313 10:15:34.923049 4930 generic.go:334] "Generic (PLEG): container finished" podID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerID="a1fad07143b6fe243430cacd47d4274b8c0136523ab869e847b3a15de19c19f9" exitCode=0 Mar 13 10:15:34 crc kubenswrapper[4930]: I0313 10:15:34.923293 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnrs5" event={"ID":"2973615f-255e-4e3a-839d-dff33bd7fb1e","Type":"ContainerDied","Data":"a1fad07143b6fe243430cacd47d4274b8c0136523ab869e847b3a15de19c19f9"} Mar 13 10:15:35 crc kubenswrapper[4930]: I0313 10:15:35.935797 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnrs5" event={"ID":"2973615f-255e-4e3a-839d-dff33bd7fb1e","Type":"ContainerStarted","Data":"5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f"} Mar 13 10:15:35 crc kubenswrapper[4930]: I0313 10:15:35.962245 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bnrs5" podStartSLOduration=2.310060472 podStartE2EDuration="9.962227157s" podCreationTimestamp="2026-03-13 10:15:26 +0000 UTC" firstStartedPulling="2026-03-13 10:15:27.836234215 +0000 UTC m=+3768.586148892" lastFinishedPulling="2026-03-13 10:15:35.4884009 +0000 UTC m=+3776.238315577" observedRunningTime="2026-03-13 10:15:35.957734845 +0000 UTC m=+3776.707649522" watchObservedRunningTime="2026-03-13 10:15:35.962227157 +0000 UTC m=+3776.712141834" Mar 13 10:15:36 crc kubenswrapper[4930]: I0313 10:15:36.907591 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:36 crc kubenswrapper[4930]: I0313 10:15:36.907813 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:15:37 crc kubenswrapper[4930]: I0313 10:15:37.954047 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bnrs5" podUID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerName="registry-server" probeResult="failure" output=< Mar 13 10:15:37 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:15:37 crc kubenswrapper[4930]: > Mar 13 10:15:37 crc kubenswrapper[4930]: I0313 10:15:37.971583 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:15:37 crc kubenswrapper[4930]: E0313 10:15:37.972159 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:15:47 crc kubenswrapper[4930]: I0313 10:15:47.964653 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bnrs5" podUID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerName="registry-server" probeResult="failure" output=< Mar 13 10:15:47 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:15:47 crc kubenswrapper[4930]: > Mar 13 10:15:50 crc kubenswrapper[4930]: I0313 10:15:50.970970 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:15:50 crc kubenswrapper[4930]: E0313 10:15:50.971770 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:15:57 crc kubenswrapper[4930]: I0313 10:15:57.967850 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bnrs5" podUID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerName="registry-server" probeResult="failure" output=< Mar 13 10:15:57 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:15:57 crc kubenswrapper[4930]: > Mar 13 10:16:00 crc kubenswrapper[4930]: I0313 10:16:00.146319 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556616-qfhn8"] Mar 13 10:16:00 crc kubenswrapper[4930]: I0313 10:16:00.148519 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556616-qfhn8" Mar 13 10:16:00 crc kubenswrapper[4930]: I0313 10:16:00.150665 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:16:00 crc kubenswrapper[4930]: I0313 10:16:00.151742 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:16:00 crc kubenswrapper[4930]: I0313 10:16:00.154981 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:16:00 crc kubenswrapper[4930]: I0313 10:16:00.158220 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556616-qfhn8"] Mar 13 10:16:00 crc kubenswrapper[4930]: I0313 10:16:00.250017 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr78t\" (UniqueName: \"kubernetes.io/projected/bf65e4e5-70cc-4d4d-a31e-010272edcf0f-kube-api-access-tr78t\") pod \"auto-csr-approver-29556616-qfhn8\" (UID: \"bf65e4e5-70cc-4d4d-a31e-010272edcf0f\") " pod="openshift-infra/auto-csr-approver-29556616-qfhn8" Mar 13 10:16:00 crc kubenswrapper[4930]: I0313 10:16:00.353180 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr78t\" (UniqueName: \"kubernetes.io/projected/bf65e4e5-70cc-4d4d-a31e-010272edcf0f-kube-api-access-tr78t\") pod \"auto-csr-approver-29556616-qfhn8\" (UID: \"bf65e4e5-70cc-4d4d-a31e-010272edcf0f\") " pod="openshift-infra/auto-csr-approver-29556616-qfhn8" Mar 13 10:16:00 crc kubenswrapper[4930]: I0313 10:16:00.371421 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr78t\" (UniqueName: \"kubernetes.io/projected/bf65e4e5-70cc-4d4d-a31e-010272edcf0f-kube-api-access-tr78t\") pod \"auto-csr-approver-29556616-qfhn8\" (UID: \"bf65e4e5-70cc-4d4d-a31e-010272edcf0f\") " pod="openshift-infra/auto-csr-approver-29556616-qfhn8" Mar 13 10:16:00 crc kubenswrapper[4930]: I0313 10:16:00.469752 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556616-qfhn8" Mar 13 10:16:01 crc kubenswrapper[4930]: I0313 10:16:01.095502 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556616-qfhn8"] Mar 13 10:16:01 crc kubenswrapper[4930]: I0313 10:16:01.249899 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556616-qfhn8" event={"ID":"bf65e4e5-70cc-4d4d-a31e-010272edcf0f","Type":"ContainerStarted","Data":"b7895566659af9ea202a91f0e919d6f71cc721cb8bd033acf2ea9802eacb0bff"} Mar 13 10:16:02 crc kubenswrapper[4930]: I0313 10:16:02.971256 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:16:02 crc kubenswrapper[4930]: E0313 10:16:02.972051 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:16:03 crc kubenswrapper[4930]: I0313 10:16:03.272016 4930 generic.go:334] "Generic (PLEG): container finished" podID="bf65e4e5-70cc-4d4d-a31e-010272edcf0f" containerID="b516251ebbf3124180be85cc79ac40c339769a56f1d99899996843200996dec8" exitCode=0 Mar 13 10:16:03 crc kubenswrapper[4930]: I0313 10:16:03.272061 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556616-qfhn8" event={"ID":"bf65e4e5-70cc-4d4d-a31e-010272edcf0f","Type":"ContainerDied","Data":"b516251ebbf3124180be85cc79ac40c339769a56f1d99899996843200996dec8"} Mar 13 10:16:04 crc kubenswrapper[4930]: I0313 10:16:04.814385 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556616-qfhn8" Mar 13 10:16:04 crc kubenswrapper[4930]: I0313 10:16:04.857215 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tr78t\" (UniqueName: \"kubernetes.io/projected/bf65e4e5-70cc-4d4d-a31e-010272edcf0f-kube-api-access-tr78t\") pod \"bf65e4e5-70cc-4d4d-a31e-010272edcf0f\" (UID: \"bf65e4e5-70cc-4d4d-a31e-010272edcf0f\") " Mar 13 10:16:04 crc kubenswrapper[4930]: I0313 10:16:04.886700 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf65e4e5-70cc-4d4d-a31e-010272edcf0f-kube-api-access-tr78t" (OuterVolumeSpecName: "kube-api-access-tr78t") pod "bf65e4e5-70cc-4d4d-a31e-010272edcf0f" (UID: "bf65e4e5-70cc-4d4d-a31e-010272edcf0f"). InnerVolumeSpecName "kube-api-access-tr78t". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:16:04 crc kubenswrapper[4930]: I0313 10:16:04.960082 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tr78t\" (UniqueName: \"kubernetes.io/projected/bf65e4e5-70cc-4d4d-a31e-010272edcf0f-kube-api-access-tr78t\") on node \"crc\" DevicePath \"\"" Mar 13 10:16:05 crc kubenswrapper[4930]: I0313 10:16:05.291818 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556616-qfhn8" event={"ID":"bf65e4e5-70cc-4d4d-a31e-010272edcf0f","Type":"ContainerDied","Data":"b7895566659af9ea202a91f0e919d6f71cc721cb8bd033acf2ea9802eacb0bff"} Mar 13 10:16:05 crc kubenswrapper[4930]: I0313 10:16:05.291862 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b7895566659af9ea202a91f0e919d6f71cc721cb8bd033acf2ea9802eacb0bff" Mar 13 10:16:05 crc kubenswrapper[4930]: I0313 10:16:05.291880 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556616-qfhn8" Mar 13 10:16:05 crc kubenswrapper[4930]: I0313 10:16:05.895704 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556610-6zvcz"] Mar 13 10:16:05 crc kubenswrapper[4930]: I0313 10:16:05.921190 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556610-6zvcz"] Mar 13 10:16:05 crc kubenswrapper[4930]: I0313 10:16:05.984346 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c020e7e-0ae4-4c30-b625-a2afc1ec5891" path="/var/lib/kubelet/pods/2c020e7e-0ae4-4c30-b625-a2afc1ec5891/volumes" Mar 13 10:16:07 crc kubenswrapper[4930]: I0313 10:16:07.063737 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:16:07 crc kubenswrapper[4930]: I0313 10:16:07.116665 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:16:07 crc kubenswrapper[4930]: I0313 10:16:07.316552 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bnrs5"] Mar 13 10:16:08 crc kubenswrapper[4930]: I0313 10:16:08.324073 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bnrs5" podUID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerName="registry-server" containerID="cri-o://5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f" gracePeriod=2 Mar 13 10:16:08 crc kubenswrapper[4930]: I0313 10:16:08.962181 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.054827 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwh42\" (UniqueName: \"kubernetes.io/projected/2973615f-255e-4e3a-839d-dff33bd7fb1e-kube-api-access-pwh42\") pod \"2973615f-255e-4e3a-839d-dff33bd7fb1e\" (UID: \"2973615f-255e-4e3a-839d-dff33bd7fb1e\") " Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.055291 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2973615f-255e-4e3a-839d-dff33bd7fb1e-catalog-content\") pod \"2973615f-255e-4e3a-839d-dff33bd7fb1e\" (UID: \"2973615f-255e-4e3a-839d-dff33bd7fb1e\") " Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.055483 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2973615f-255e-4e3a-839d-dff33bd7fb1e-utilities\") pod \"2973615f-255e-4e3a-839d-dff33bd7fb1e\" (UID: \"2973615f-255e-4e3a-839d-dff33bd7fb1e\") " Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.056129 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2973615f-255e-4e3a-839d-dff33bd7fb1e-utilities" (OuterVolumeSpecName: "utilities") pod "2973615f-255e-4e3a-839d-dff33bd7fb1e" (UID: "2973615f-255e-4e3a-839d-dff33bd7fb1e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.056486 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2973615f-255e-4e3a-839d-dff33bd7fb1e-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.061447 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2973615f-255e-4e3a-839d-dff33bd7fb1e-kube-api-access-pwh42" (OuterVolumeSpecName: "kube-api-access-pwh42") pod "2973615f-255e-4e3a-839d-dff33bd7fb1e" (UID: "2973615f-255e-4e3a-839d-dff33bd7fb1e"). InnerVolumeSpecName "kube-api-access-pwh42". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.158888 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwh42\" (UniqueName: \"kubernetes.io/projected/2973615f-255e-4e3a-839d-dff33bd7fb1e-kube-api-access-pwh42\") on node \"crc\" DevicePath \"\"" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.197116 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2973615f-255e-4e3a-839d-dff33bd7fb1e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2973615f-255e-4e3a-839d-dff33bd7fb1e" (UID: "2973615f-255e-4e3a-839d-dff33bd7fb1e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.261867 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2973615f-255e-4e3a-839d-dff33bd7fb1e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.339233 4930 generic.go:334] "Generic (PLEG): container finished" podID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerID="5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f" exitCode=0 Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.339316 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnrs5" event={"ID":"2973615f-255e-4e3a-839d-dff33bd7fb1e","Type":"ContainerDied","Data":"5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f"} Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.340298 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bnrs5" event={"ID":"2973615f-255e-4e3a-839d-dff33bd7fb1e","Type":"ContainerDied","Data":"e2a1c2eb9dd663d4282c92162e4c31e0d663ff53bb40e7893136efff4e1e8ae5"} Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.339359 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bnrs5" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.340366 4930 scope.go:117] "RemoveContainer" containerID="5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.363830 4930 scope.go:117] "RemoveContainer" containerID="a1fad07143b6fe243430cacd47d4274b8c0136523ab869e847b3a15de19c19f9" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.398293 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bnrs5"] Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.426642 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bnrs5"] Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.426945 4930 scope.go:117] "RemoveContainer" containerID="b1aefee2a21791715c6bb68a14631bfb5b031b5d5af9989b61f48d491ea68012" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.465317 4930 scope.go:117] "RemoveContainer" containerID="5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f" Mar 13 10:16:09 crc kubenswrapper[4930]: E0313 10:16:09.466104 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f\": container with ID starting with 5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f not found: ID does not exist" containerID="5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.466137 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f"} err="failed to get container status \"5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f\": rpc error: code = NotFound desc = could not find container \"5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f\": container with ID starting with 5d90206f52b832521b7b4ad614fc2c09b700f2fb17103c4245019734cc05f49f not found: ID does not exist" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.466157 4930 scope.go:117] "RemoveContainer" containerID="a1fad07143b6fe243430cacd47d4274b8c0136523ab869e847b3a15de19c19f9" Mar 13 10:16:09 crc kubenswrapper[4930]: E0313 10:16:09.466490 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1fad07143b6fe243430cacd47d4274b8c0136523ab869e847b3a15de19c19f9\": container with ID starting with a1fad07143b6fe243430cacd47d4274b8c0136523ab869e847b3a15de19c19f9 not found: ID does not exist" containerID="a1fad07143b6fe243430cacd47d4274b8c0136523ab869e847b3a15de19c19f9" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.466585 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1fad07143b6fe243430cacd47d4274b8c0136523ab869e847b3a15de19c19f9"} err="failed to get container status \"a1fad07143b6fe243430cacd47d4274b8c0136523ab869e847b3a15de19c19f9\": rpc error: code = NotFound desc = could not find container \"a1fad07143b6fe243430cacd47d4274b8c0136523ab869e847b3a15de19c19f9\": container with ID starting with a1fad07143b6fe243430cacd47d4274b8c0136523ab869e847b3a15de19c19f9 not found: ID does not exist" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.466777 4930 scope.go:117] "RemoveContainer" containerID="b1aefee2a21791715c6bb68a14631bfb5b031b5d5af9989b61f48d491ea68012" Mar 13 10:16:09 crc kubenswrapper[4930]: E0313 10:16:09.467269 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b1aefee2a21791715c6bb68a14631bfb5b031b5d5af9989b61f48d491ea68012\": container with ID starting with b1aefee2a21791715c6bb68a14631bfb5b031b5d5af9989b61f48d491ea68012 not found: ID does not exist" containerID="b1aefee2a21791715c6bb68a14631bfb5b031b5d5af9989b61f48d491ea68012" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.467320 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b1aefee2a21791715c6bb68a14631bfb5b031b5d5af9989b61f48d491ea68012"} err="failed to get container status \"b1aefee2a21791715c6bb68a14631bfb5b031b5d5af9989b61f48d491ea68012\": rpc error: code = NotFound desc = could not find container \"b1aefee2a21791715c6bb68a14631bfb5b031b5d5af9989b61f48d491ea68012\": container with ID starting with b1aefee2a21791715c6bb68a14631bfb5b031b5d5af9989b61f48d491ea68012 not found: ID does not exist" Mar 13 10:16:09 crc kubenswrapper[4930]: I0313 10:16:09.983582 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2973615f-255e-4e3a-839d-dff33bd7fb1e" path="/var/lib/kubelet/pods/2973615f-255e-4e3a-839d-dff33bd7fb1e/volumes" Mar 13 10:16:16 crc kubenswrapper[4930]: I0313 10:16:16.970663 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:16:16 crc kubenswrapper[4930]: E0313 10:16:16.971666 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:16:30 crc kubenswrapper[4930]: I0313 10:16:30.971502 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:16:30 crc kubenswrapper[4930]: E0313 10:16:30.972249 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:16:32 crc kubenswrapper[4930]: I0313 10:16:32.699734 4930 scope.go:117] "RemoveContainer" containerID="7edfd81c58a94c68ba605930c9cae1138728751b1f581f0228adff0db7b05a6d" Mar 13 10:16:44 crc kubenswrapper[4930]: I0313 10:16:44.971597 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:16:44 crc kubenswrapper[4930]: E0313 10:16:44.972597 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:16:59 crc kubenswrapper[4930]: I0313 10:16:59.971185 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:16:59 crc kubenswrapper[4930]: E0313 10:16:59.971979 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:17:11 crc kubenswrapper[4930]: I0313 10:17:11.970816 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:17:11 crc kubenswrapper[4930]: E0313 10:17:11.971672 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:17:24 crc kubenswrapper[4930]: I0313 10:17:24.971011 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:17:24 crc kubenswrapper[4930]: E0313 10:17:24.971926 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:17:32 crc kubenswrapper[4930]: I0313 10:17:32.169072 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" podUID="046b75aa-f3a0-467d-8624-6a82aa748b32" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:17:32 crc kubenswrapper[4930]: I0313 10:17:32.182301 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" podUID="47a82845-cf9f-4fc8-a35f-4548a9c7127e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": dial tcp 10.217.0.114:8081: i/o timeout (Client.Timeout exceeded while awaiting headers)" Mar 13 10:17:32 crc kubenswrapper[4930]: I0313 10:17:32.261738 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" podUID="4f1e0b58-9978-423f-84bc-f974dc6347b4" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:17:34 crc kubenswrapper[4930]: E0313 10:17:34.919077 4930 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.238:36460->38.102.83.238:33987: write tcp 38.102.83.238:36460->38.102.83.238:33987: write: broken pipe Mar 13 10:17:38 crc kubenswrapper[4930]: I0313 10:17:38.971388 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:17:38 crc kubenswrapper[4930]: E0313 10:17:38.972258 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:17:52 crc kubenswrapper[4930]: I0313 10:17:51.971399 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:17:52 crc kubenswrapper[4930]: E0313 10:17:51.972522 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.147673 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556618-bh7lw"] Mar 13 10:18:00 crc kubenswrapper[4930]: E0313 10:18:00.148897 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerName="extract-content" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.148916 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerName="extract-content" Mar 13 10:18:00 crc kubenswrapper[4930]: E0313 10:18:00.148932 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerName="registry-server" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.148938 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerName="registry-server" Mar 13 10:18:00 crc kubenswrapper[4930]: E0313 10:18:00.148953 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf65e4e5-70cc-4d4d-a31e-010272edcf0f" containerName="oc" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.148961 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf65e4e5-70cc-4d4d-a31e-010272edcf0f" containerName="oc" Mar 13 10:18:00 crc kubenswrapper[4930]: E0313 10:18:00.148981 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerName="extract-utilities" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.148987 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerName="extract-utilities" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.149257 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf65e4e5-70cc-4d4d-a31e-010272edcf0f" containerName="oc" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.149277 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="2973615f-255e-4e3a-839d-dff33bd7fb1e" containerName="registry-server" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.150226 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556618-bh7lw" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.154169 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.154624 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.154951 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.173553 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556618-bh7lw"] Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.212344 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8swzx\" (UniqueName: \"kubernetes.io/projected/5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1-kube-api-access-8swzx\") pod \"auto-csr-approver-29556618-bh7lw\" (UID: \"5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1\") " pod="openshift-infra/auto-csr-approver-29556618-bh7lw" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.314662 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8swzx\" (UniqueName: \"kubernetes.io/projected/5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1-kube-api-access-8swzx\") pod \"auto-csr-approver-29556618-bh7lw\" (UID: \"5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1\") " pod="openshift-infra/auto-csr-approver-29556618-bh7lw" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.339291 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8swzx\" (UniqueName: \"kubernetes.io/projected/5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1-kube-api-access-8swzx\") pod \"auto-csr-approver-29556618-bh7lw\" (UID: \"5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1\") " pod="openshift-infra/auto-csr-approver-29556618-bh7lw" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.471081 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556618-bh7lw" Mar 13 10:18:00 crc kubenswrapper[4930]: I0313 10:18:00.932042 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556618-bh7lw"] Mar 13 10:18:01 crc kubenswrapper[4930]: I0313 10:18:01.626828 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556618-bh7lw" event={"ID":"5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1","Type":"ContainerStarted","Data":"0fbfbdbd45f69d741141ed1f032072aa17111cdcb1515685a53735476731a31c"} Mar 13 10:18:03 crc kubenswrapper[4930]: I0313 10:18:03.647865 4930 generic.go:334] "Generic (PLEG): container finished" podID="5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1" containerID="b1e995e5f882c69d87520b16aba7b659e90c0b3e164b93342f09038af30ac081" exitCode=0 Mar 13 10:18:03 crc kubenswrapper[4930]: I0313 10:18:03.647914 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556618-bh7lw" event={"ID":"5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1","Type":"ContainerDied","Data":"b1e995e5f882c69d87520b16aba7b659e90c0b3e164b93342f09038af30ac081"} Mar 13 10:18:05 crc kubenswrapper[4930]: I0313 10:18:05.082704 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556618-bh7lw" Mar 13 10:18:05 crc kubenswrapper[4930]: I0313 10:18:05.136082 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8swzx\" (UniqueName: \"kubernetes.io/projected/5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1-kube-api-access-8swzx\") pod \"5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1\" (UID: \"5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1\") " Mar 13 10:18:05 crc kubenswrapper[4930]: I0313 10:18:05.159857 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1-kube-api-access-8swzx" (OuterVolumeSpecName: "kube-api-access-8swzx") pod "5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1" (UID: "5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1"). InnerVolumeSpecName "kube-api-access-8swzx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:18:05 crc kubenswrapper[4930]: I0313 10:18:05.240810 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8swzx\" (UniqueName: \"kubernetes.io/projected/5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1-kube-api-access-8swzx\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:05 crc kubenswrapper[4930]: I0313 10:18:05.671885 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556618-bh7lw" event={"ID":"5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1","Type":"ContainerDied","Data":"0fbfbdbd45f69d741141ed1f032072aa17111cdcb1515685a53735476731a31c"} Mar 13 10:18:05 crc kubenswrapper[4930]: I0313 10:18:05.671926 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0fbfbdbd45f69d741141ed1f032072aa17111cdcb1515685a53735476731a31c" Mar 13 10:18:05 crc kubenswrapper[4930]: I0313 10:18:05.671940 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556618-bh7lw" Mar 13 10:18:05 crc kubenswrapper[4930]: I0313 10:18:05.971657 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:18:05 crc kubenswrapper[4930]: E0313 10:18:05.972066 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:18:06 crc kubenswrapper[4930]: I0313 10:18:06.160223 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556612-88shc"] Mar 13 10:18:06 crc kubenswrapper[4930]: I0313 10:18:06.174726 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556612-88shc"] Mar 13 10:18:07 crc kubenswrapper[4930]: I0313 10:18:07.987379 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afdb2c21-1f79-4d0f-a99d-d72b782ccbf5" path="/var/lib/kubelet/pods/afdb2c21-1f79-4d0f-a99d-d72b782ccbf5/volumes" Mar 13 10:18:18 crc kubenswrapper[4930]: I0313 10:18:18.971219 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:18:19 crc kubenswrapper[4930]: I0313 10:18:19.842299 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"a2dd0e422956ad3107b66f8a5d96648c6e90c0a7f1dd96ff6dcb9a41253b535b"} Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.162076 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9cbzs"] Mar 13 10:18:20 crc kubenswrapper[4930]: E0313 10:18:20.162959 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1" containerName="oc" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.162973 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1" containerName="oc" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.163255 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1" containerName="oc" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.165985 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.181637 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9cbzs"] Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.264138 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-catalog-content\") pod \"community-operators-9cbzs\" (UID: \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\") " pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.264235 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-utilities\") pod \"community-operators-9cbzs\" (UID: \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\") " pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.264534 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txlcs\" (UniqueName: \"kubernetes.io/projected/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-kube-api-access-txlcs\") pod \"community-operators-9cbzs\" (UID: \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\") " pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.367621 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txlcs\" (UniqueName: \"kubernetes.io/projected/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-kube-api-access-txlcs\") pod \"community-operators-9cbzs\" (UID: \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\") " pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.367731 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-catalog-content\") pod \"community-operators-9cbzs\" (UID: \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\") " pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.367782 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-utilities\") pod \"community-operators-9cbzs\" (UID: \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\") " pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.368194 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-catalog-content\") pod \"community-operators-9cbzs\" (UID: \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\") " pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.368252 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-utilities\") pod \"community-operators-9cbzs\" (UID: \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\") " pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.390479 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txlcs\" (UniqueName: \"kubernetes.io/projected/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-kube-api-access-txlcs\") pod \"community-operators-9cbzs\" (UID: \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\") " pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:20 crc kubenswrapper[4930]: I0313 10:18:20.518731 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:21 crc kubenswrapper[4930]: I0313 10:18:21.171052 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9cbzs"] Mar 13 10:18:21 crc kubenswrapper[4930]: W0313 10:18:21.172684 4930 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod25c7fd0f_1bdb_4a93_b4b5_1272d09ca070.slice/crio-9c8b7cb92f350e48c399df5ba6617797466a6c558213247e507f4d2437571697 WatchSource:0}: Error finding container 9c8b7cb92f350e48c399df5ba6617797466a6c558213247e507f4d2437571697: Status 404 returned error can't find the container with id 9c8b7cb92f350e48c399df5ba6617797466a6c558213247e507f4d2437571697 Mar 13 10:18:21 crc kubenswrapper[4930]: I0313 10:18:21.876154 4930 generic.go:334] "Generic (PLEG): container finished" podID="25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" containerID="70416706c3b859bc93720b594e997216413bb6ce4e089e37ec43c97d31352277" exitCode=0 Mar 13 10:18:21 crc kubenswrapper[4930]: I0313 10:18:21.876265 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9cbzs" event={"ID":"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070","Type":"ContainerDied","Data":"70416706c3b859bc93720b594e997216413bb6ce4e089e37ec43c97d31352277"} Mar 13 10:18:21 crc kubenswrapper[4930]: I0313 10:18:21.876600 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9cbzs" event={"ID":"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070","Type":"ContainerStarted","Data":"9c8b7cb92f350e48c399df5ba6617797466a6c558213247e507f4d2437571697"} Mar 13 10:18:23 crc kubenswrapper[4930]: I0313 10:18:23.985786 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9cbzs" event={"ID":"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070","Type":"ContainerStarted","Data":"ce541eafc8fc2c4c4a33ab7b847a9ac2edf513414d4e63ae1a37ba9091ef062b"} Mar 13 10:18:25 crc kubenswrapper[4930]: I0313 10:18:25.997223 4930 generic.go:334] "Generic (PLEG): container finished" podID="25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" containerID="ce541eafc8fc2c4c4a33ab7b847a9ac2edf513414d4e63ae1a37ba9091ef062b" exitCode=0 Mar 13 10:18:25 crc kubenswrapper[4930]: I0313 10:18:25.997334 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9cbzs" event={"ID":"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070","Type":"ContainerDied","Data":"ce541eafc8fc2c4c4a33ab7b847a9ac2edf513414d4e63ae1a37ba9091ef062b"} Mar 13 10:18:27 crc kubenswrapper[4930]: I0313 10:18:27.027610 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9cbzs" event={"ID":"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070","Type":"ContainerStarted","Data":"0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7"} Mar 13 10:18:27 crc kubenswrapper[4930]: I0313 10:18:27.053608 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9cbzs" podStartSLOduration=2.387717206 podStartE2EDuration="7.05357916s" podCreationTimestamp="2026-03-13 10:18:20 +0000 UTC" firstStartedPulling="2026-03-13 10:18:21.881864524 +0000 UTC m=+3942.631779201" lastFinishedPulling="2026-03-13 10:18:26.547726468 +0000 UTC m=+3947.297641155" observedRunningTime="2026-03-13 10:18:27.052606475 +0000 UTC m=+3947.802521162" watchObservedRunningTime="2026-03-13 10:18:27.05357916 +0000 UTC m=+3947.803493837" Mar 13 10:18:30 crc kubenswrapper[4930]: I0313 10:18:30.519154 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:30 crc kubenswrapper[4930]: I0313 10:18:30.520349 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:30 crc kubenswrapper[4930]: I0313 10:18:30.569040 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:31 crc kubenswrapper[4930]: I0313 10:18:31.143290 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:31 crc kubenswrapper[4930]: I0313 10:18:31.207901 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9cbzs"] Mar 13 10:18:32 crc kubenswrapper[4930]: I0313 10:18:32.826601 4930 scope.go:117] "RemoveContainer" containerID="1592e8677b50e62c256a8c8f25273e374912dcbfad982e344e17d3bdd694fe61" Mar 13 10:18:33 crc kubenswrapper[4930]: I0313 10:18:33.111590 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9cbzs" podUID="25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" containerName="registry-server" containerID="cri-o://0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7" gracePeriod=2 Mar 13 10:18:33 crc kubenswrapper[4930]: I0313 10:18:33.689055 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:33 crc kubenswrapper[4930]: I0313 10:18:33.775951 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-catalog-content\") pod \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\" (UID: \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\") " Mar 13 10:18:33 crc kubenswrapper[4930]: I0313 10:18:33.777058 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txlcs\" (UniqueName: \"kubernetes.io/projected/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-kube-api-access-txlcs\") pod \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\" (UID: \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\") " Mar 13 10:18:33 crc kubenswrapper[4930]: I0313 10:18:33.777130 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-utilities\") pod \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\" (UID: \"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070\") " Mar 13 10:18:33 crc kubenswrapper[4930]: I0313 10:18:33.778763 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-utilities" (OuterVolumeSpecName: "utilities") pod "25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" (UID: "25c7fd0f-1bdb-4a93-b4b5-1272d09ca070"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:18:33 crc kubenswrapper[4930]: I0313 10:18:33.779719 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:33 crc kubenswrapper[4930]: I0313 10:18:33.802283 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-kube-api-access-txlcs" (OuterVolumeSpecName: "kube-api-access-txlcs") pod "25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" (UID: "25c7fd0f-1bdb-4a93-b4b5-1272d09ca070"). InnerVolumeSpecName "kube-api-access-txlcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:18:33 crc kubenswrapper[4930]: I0313 10:18:33.848184 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" (UID: "25c7fd0f-1bdb-4a93-b4b5-1272d09ca070"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:18:33 crc kubenswrapper[4930]: I0313 10:18:33.882791 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:33 crc kubenswrapper[4930]: I0313 10:18:33.882828 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txlcs\" (UniqueName: \"kubernetes.io/projected/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070-kube-api-access-txlcs\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.125219 4930 generic.go:334] "Generic (PLEG): container finished" podID="25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" containerID="0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7" exitCode=0 Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.125273 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9cbzs" event={"ID":"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070","Type":"ContainerDied","Data":"0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7"} Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.125309 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9cbzs" event={"ID":"25c7fd0f-1bdb-4a93-b4b5-1272d09ca070","Type":"ContainerDied","Data":"9c8b7cb92f350e48c399df5ba6617797466a6c558213247e507f4d2437571697"} Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.125335 4930 scope.go:117] "RemoveContainer" containerID="0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7" Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.125561 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9cbzs" Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.159702 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9cbzs"] Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.168501 4930 scope.go:117] "RemoveContainer" containerID="ce541eafc8fc2c4c4a33ab7b847a9ac2edf513414d4e63ae1a37ba9091ef062b" Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.178872 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9cbzs"] Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.204686 4930 scope.go:117] "RemoveContainer" containerID="70416706c3b859bc93720b594e997216413bb6ce4e089e37ec43c97d31352277" Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.263735 4930 scope.go:117] "RemoveContainer" containerID="0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7" Mar 13 10:18:34 crc kubenswrapper[4930]: E0313 10:18:34.264153 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7\": container with ID starting with 0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7 not found: ID does not exist" containerID="0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7" Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.264193 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7"} err="failed to get container status \"0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7\": rpc error: code = NotFound desc = could not find container \"0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7\": container with ID starting with 0702838840a1a7bdbc3b79696d9c2659407785e52b6784e6397a3f3a5b7a5ab7 not found: ID does not exist" Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.264219 4930 scope.go:117] "RemoveContainer" containerID="ce541eafc8fc2c4c4a33ab7b847a9ac2edf513414d4e63ae1a37ba9091ef062b" Mar 13 10:18:34 crc kubenswrapper[4930]: E0313 10:18:34.264530 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce541eafc8fc2c4c4a33ab7b847a9ac2edf513414d4e63ae1a37ba9091ef062b\": container with ID starting with ce541eafc8fc2c4c4a33ab7b847a9ac2edf513414d4e63ae1a37ba9091ef062b not found: ID does not exist" containerID="ce541eafc8fc2c4c4a33ab7b847a9ac2edf513414d4e63ae1a37ba9091ef062b" Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.264565 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce541eafc8fc2c4c4a33ab7b847a9ac2edf513414d4e63ae1a37ba9091ef062b"} err="failed to get container status \"ce541eafc8fc2c4c4a33ab7b847a9ac2edf513414d4e63ae1a37ba9091ef062b\": rpc error: code = NotFound desc = could not find container \"ce541eafc8fc2c4c4a33ab7b847a9ac2edf513414d4e63ae1a37ba9091ef062b\": container with ID starting with ce541eafc8fc2c4c4a33ab7b847a9ac2edf513414d4e63ae1a37ba9091ef062b not found: ID does not exist" Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.264585 4930 scope.go:117] "RemoveContainer" containerID="70416706c3b859bc93720b594e997216413bb6ce4e089e37ec43c97d31352277" Mar 13 10:18:34 crc kubenswrapper[4930]: E0313 10:18:34.264854 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"70416706c3b859bc93720b594e997216413bb6ce4e089e37ec43c97d31352277\": container with ID starting with 70416706c3b859bc93720b594e997216413bb6ce4e089e37ec43c97d31352277 not found: ID does not exist" containerID="70416706c3b859bc93720b594e997216413bb6ce4e089e37ec43c97d31352277" Mar 13 10:18:34 crc kubenswrapper[4930]: I0313 10:18:34.264876 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"70416706c3b859bc93720b594e997216413bb6ce4e089e37ec43c97d31352277"} err="failed to get container status \"70416706c3b859bc93720b594e997216413bb6ce4e089e37ec43c97d31352277\": rpc error: code = NotFound desc = could not find container \"70416706c3b859bc93720b594e997216413bb6ce4e089e37ec43c97d31352277\": container with ID starting with 70416706c3b859bc93720b594e997216413bb6ce4e089e37ec43c97d31352277 not found: ID does not exist" Mar 13 10:18:35 crc kubenswrapper[4930]: I0313 10:18:35.988862 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" path="/var/lib/kubelet/pods/25c7fd0f-1bdb-4a93-b4b5-1272d09ca070/volumes" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.321875 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8jgv6"] Mar 13 10:18:39 crc kubenswrapper[4930]: E0313 10:18:39.322854 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" containerName="extract-content" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.322872 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" containerName="extract-content" Mar 13 10:18:39 crc kubenswrapper[4930]: E0313 10:18:39.322884 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" containerName="extract-utilities" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.322891 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" containerName="extract-utilities" Mar 13 10:18:39 crc kubenswrapper[4930]: E0313 10:18:39.322922 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" containerName="registry-server" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.322931 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" containerName="registry-server" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.323209 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="25c7fd0f-1bdb-4a93-b4b5-1272d09ca070" containerName="registry-server" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.325709 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.337679 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8jgv6"] Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.421983 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smfnh\" (UniqueName: \"kubernetes.io/projected/a7181af2-26ac-48dd-988a-0b606b282234-kube-api-access-smfnh\") pod \"redhat-marketplace-8jgv6\" (UID: \"a7181af2-26ac-48dd-988a-0b606b282234\") " pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.422052 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7181af2-26ac-48dd-988a-0b606b282234-catalog-content\") pod \"redhat-marketplace-8jgv6\" (UID: \"a7181af2-26ac-48dd-988a-0b606b282234\") " pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.422088 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7181af2-26ac-48dd-988a-0b606b282234-utilities\") pod \"redhat-marketplace-8jgv6\" (UID: \"a7181af2-26ac-48dd-988a-0b606b282234\") " pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.524944 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7181af2-26ac-48dd-988a-0b606b282234-utilities\") pod \"redhat-marketplace-8jgv6\" (UID: \"a7181af2-26ac-48dd-988a-0b606b282234\") " pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.525237 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smfnh\" (UniqueName: \"kubernetes.io/projected/a7181af2-26ac-48dd-988a-0b606b282234-kube-api-access-smfnh\") pod \"redhat-marketplace-8jgv6\" (UID: \"a7181af2-26ac-48dd-988a-0b606b282234\") " pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.525294 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7181af2-26ac-48dd-988a-0b606b282234-catalog-content\") pod \"redhat-marketplace-8jgv6\" (UID: \"a7181af2-26ac-48dd-988a-0b606b282234\") " pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.525569 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7181af2-26ac-48dd-988a-0b606b282234-utilities\") pod \"redhat-marketplace-8jgv6\" (UID: \"a7181af2-26ac-48dd-988a-0b606b282234\") " pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.525642 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7181af2-26ac-48dd-988a-0b606b282234-catalog-content\") pod \"redhat-marketplace-8jgv6\" (UID: \"a7181af2-26ac-48dd-988a-0b606b282234\") " pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.559345 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smfnh\" (UniqueName: \"kubernetes.io/projected/a7181af2-26ac-48dd-988a-0b606b282234-kube-api-access-smfnh\") pod \"redhat-marketplace-8jgv6\" (UID: \"a7181af2-26ac-48dd-988a-0b606b282234\") " pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:39 crc kubenswrapper[4930]: I0313 10:18:39.708040 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:40 crc kubenswrapper[4930]: I0313 10:18:40.212156 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8jgv6"] Mar 13 10:18:41 crc kubenswrapper[4930]: I0313 10:18:41.205024 4930 generic.go:334] "Generic (PLEG): container finished" podID="a7181af2-26ac-48dd-988a-0b606b282234" containerID="bfe33c28e57d18b01bfd1f2768ccaf86e375c64fb9648e4692e96794b2296939" exitCode=0 Mar 13 10:18:41 crc kubenswrapper[4930]: I0313 10:18:41.206230 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8jgv6" event={"ID":"a7181af2-26ac-48dd-988a-0b606b282234","Type":"ContainerDied","Data":"bfe33c28e57d18b01bfd1f2768ccaf86e375c64fb9648e4692e96794b2296939"} Mar 13 10:18:41 crc kubenswrapper[4930]: I0313 10:18:41.206316 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8jgv6" event={"ID":"a7181af2-26ac-48dd-988a-0b606b282234","Type":"ContainerStarted","Data":"8923fbfee534244da9b6911789971cce293d04db377bdb304e66358ce071ec0e"} Mar 13 10:18:43 crc kubenswrapper[4930]: I0313 10:18:43.225814 4930 generic.go:334] "Generic (PLEG): container finished" podID="a7181af2-26ac-48dd-988a-0b606b282234" containerID="68590c93fb12f830b8294b61805b69ef9aa7c6ab11a4b048cab3d42d6a705c63" exitCode=0 Mar 13 10:18:43 crc kubenswrapper[4930]: I0313 10:18:43.226028 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8jgv6" event={"ID":"a7181af2-26ac-48dd-988a-0b606b282234","Type":"ContainerDied","Data":"68590c93fb12f830b8294b61805b69ef9aa7c6ab11a4b048cab3d42d6a705c63"} Mar 13 10:18:44 crc kubenswrapper[4930]: I0313 10:18:44.237496 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8jgv6" event={"ID":"a7181af2-26ac-48dd-988a-0b606b282234","Type":"ContainerStarted","Data":"bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197"} Mar 13 10:18:44 crc kubenswrapper[4930]: I0313 10:18:44.262794 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8jgv6" podStartSLOduration=2.776816332 podStartE2EDuration="5.262770521s" podCreationTimestamp="2026-03-13 10:18:39 +0000 UTC" firstStartedPulling="2026-03-13 10:18:41.207874473 +0000 UTC m=+3961.957789150" lastFinishedPulling="2026-03-13 10:18:43.693828662 +0000 UTC m=+3964.443743339" observedRunningTime="2026-03-13 10:18:44.254458272 +0000 UTC m=+3965.004372969" watchObservedRunningTime="2026-03-13 10:18:44.262770521 +0000 UTC m=+3965.012685208" Mar 13 10:18:49 crc kubenswrapper[4930]: I0313 10:18:49.708646 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:49 crc kubenswrapper[4930]: I0313 10:18:49.709410 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:49 crc kubenswrapper[4930]: I0313 10:18:49.776184 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:50 crc kubenswrapper[4930]: I0313 10:18:50.357984 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:51 crc kubenswrapper[4930]: I0313 10:18:51.027223 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8jgv6"] Mar 13 10:18:52 crc kubenswrapper[4930]: I0313 10:18:52.409549 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8jgv6" podUID="a7181af2-26ac-48dd-988a-0b606b282234" containerName="registry-server" containerID="cri-o://bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197" gracePeriod=2 Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.015524 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.199896 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-smfnh\" (UniqueName: \"kubernetes.io/projected/a7181af2-26ac-48dd-988a-0b606b282234-kube-api-access-smfnh\") pod \"a7181af2-26ac-48dd-988a-0b606b282234\" (UID: \"a7181af2-26ac-48dd-988a-0b606b282234\") " Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.199997 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7181af2-26ac-48dd-988a-0b606b282234-utilities\") pod \"a7181af2-26ac-48dd-988a-0b606b282234\" (UID: \"a7181af2-26ac-48dd-988a-0b606b282234\") " Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.200140 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7181af2-26ac-48dd-988a-0b606b282234-catalog-content\") pod \"a7181af2-26ac-48dd-988a-0b606b282234\" (UID: \"a7181af2-26ac-48dd-988a-0b606b282234\") " Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.201104 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7181af2-26ac-48dd-988a-0b606b282234-utilities" (OuterVolumeSpecName: "utilities") pod "a7181af2-26ac-48dd-988a-0b606b282234" (UID: "a7181af2-26ac-48dd-988a-0b606b282234"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.207685 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7181af2-26ac-48dd-988a-0b606b282234-kube-api-access-smfnh" (OuterVolumeSpecName: "kube-api-access-smfnh") pod "a7181af2-26ac-48dd-988a-0b606b282234" (UID: "a7181af2-26ac-48dd-988a-0b606b282234"). InnerVolumeSpecName "kube-api-access-smfnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.231633 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7181af2-26ac-48dd-988a-0b606b282234-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7181af2-26ac-48dd-988a-0b606b282234" (UID: "a7181af2-26ac-48dd-988a-0b606b282234"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.303672 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-smfnh\" (UniqueName: \"kubernetes.io/projected/a7181af2-26ac-48dd-988a-0b606b282234-kube-api-access-smfnh\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.303729 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7181af2-26ac-48dd-988a-0b606b282234-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.303750 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7181af2-26ac-48dd-988a-0b606b282234-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.421555 4930 generic.go:334] "Generic (PLEG): container finished" podID="a7181af2-26ac-48dd-988a-0b606b282234" containerID="bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197" exitCode=0 Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.421592 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8jgv6" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.421613 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8jgv6" event={"ID":"a7181af2-26ac-48dd-988a-0b606b282234","Type":"ContainerDied","Data":"bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197"} Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.421646 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8jgv6" event={"ID":"a7181af2-26ac-48dd-988a-0b606b282234","Type":"ContainerDied","Data":"8923fbfee534244da9b6911789971cce293d04db377bdb304e66358ce071ec0e"} Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.421667 4930 scope.go:117] "RemoveContainer" containerID="bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.449897 4930 scope.go:117] "RemoveContainer" containerID="68590c93fb12f830b8294b61805b69ef9aa7c6ab11a4b048cab3d42d6a705c63" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.455510 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8jgv6"] Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.466939 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8jgv6"] Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.479336 4930 scope.go:117] "RemoveContainer" containerID="bfe33c28e57d18b01bfd1f2768ccaf86e375c64fb9648e4692e96794b2296939" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.537007 4930 scope.go:117] "RemoveContainer" containerID="bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197" Mar 13 10:18:53 crc kubenswrapper[4930]: E0313 10:18:53.537455 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197\": container with ID starting with bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197 not found: ID does not exist" containerID="bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.537496 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197"} err="failed to get container status \"bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197\": rpc error: code = NotFound desc = could not find container \"bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197\": container with ID starting with bed6161ab11a3e4dd585b0e35265b4c85bdee043945641da1aa9612d3a707197 not found: ID does not exist" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.537524 4930 scope.go:117] "RemoveContainer" containerID="68590c93fb12f830b8294b61805b69ef9aa7c6ab11a4b048cab3d42d6a705c63" Mar 13 10:18:53 crc kubenswrapper[4930]: E0313 10:18:53.537919 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68590c93fb12f830b8294b61805b69ef9aa7c6ab11a4b048cab3d42d6a705c63\": container with ID starting with 68590c93fb12f830b8294b61805b69ef9aa7c6ab11a4b048cab3d42d6a705c63 not found: ID does not exist" containerID="68590c93fb12f830b8294b61805b69ef9aa7c6ab11a4b048cab3d42d6a705c63" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.537961 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68590c93fb12f830b8294b61805b69ef9aa7c6ab11a4b048cab3d42d6a705c63"} err="failed to get container status \"68590c93fb12f830b8294b61805b69ef9aa7c6ab11a4b048cab3d42d6a705c63\": rpc error: code = NotFound desc = could not find container \"68590c93fb12f830b8294b61805b69ef9aa7c6ab11a4b048cab3d42d6a705c63\": container with ID starting with 68590c93fb12f830b8294b61805b69ef9aa7c6ab11a4b048cab3d42d6a705c63 not found: ID does not exist" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.537987 4930 scope.go:117] "RemoveContainer" containerID="bfe33c28e57d18b01bfd1f2768ccaf86e375c64fb9648e4692e96794b2296939" Mar 13 10:18:53 crc kubenswrapper[4930]: E0313 10:18:53.538300 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfe33c28e57d18b01bfd1f2768ccaf86e375c64fb9648e4692e96794b2296939\": container with ID starting with bfe33c28e57d18b01bfd1f2768ccaf86e375c64fb9648e4692e96794b2296939 not found: ID does not exist" containerID="bfe33c28e57d18b01bfd1f2768ccaf86e375c64fb9648e4692e96794b2296939" Mar 13 10:18:53 crc kubenswrapper[4930]: I0313 10:18:53.538334 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfe33c28e57d18b01bfd1f2768ccaf86e375c64fb9648e4692e96794b2296939"} err="failed to get container status \"bfe33c28e57d18b01bfd1f2768ccaf86e375c64fb9648e4692e96794b2296939\": rpc error: code = NotFound desc = could not find container \"bfe33c28e57d18b01bfd1f2768ccaf86e375c64fb9648e4692e96794b2296939\": container with ID starting with bfe33c28e57d18b01bfd1f2768ccaf86e375c64fb9648e4692e96794b2296939 not found: ID does not exist" Mar 13 10:18:54 crc kubenswrapper[4930]: I0313 10:18:54.004805 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7181af2-26ac-48dd-988a-0b606b282234" path="/var/lib/kubelet/pods/a7181af2-26ac-48dd-988a-0b606b282234/volumes" Mar 13 10:19:35 crc kubenswrapper[4930]: I0313 10:19:35.855472 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4j8jm"] Mar 13 10:19:35 crc kubenswrapper[4930]: E0313 10:19:35.856716 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7181af2-26ac-48dd-988a-0b606b282234" containerName="extract-utilities" Mar 13 10:19:35 crc kubenswrapper[4930]: I0313 10:19:35.856736 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7181af2-26ac-48dd-988a-0b606b282234" containerName="extract-utilities" Mar 13 10:19:35 crc kubenswrapper[4930]: E0313 10:19:35.856789 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7181af2-26ac-48dd-988a-0b606b282234" containerName="extract-content" Mar 13 10:19:35 crc kubenswrapper[4930]: I0313 10:19:35.856799 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7181af2-26ac-48dd-988a-0b606b282234" containerName="extract-content" Mar 13 10:19:35 crc kubenswrapper[4930]: E0313 10:19:35.856833 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7181af2-26ac-48dd-988a-0b606b282234" containerName="registry-server" Mar 13 10:19:35 crc kubenswrapper[4930]: I0313 10:19:35.856841 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7181af2-26ac-48dd-988a-0b606b282234" containerName="registry-server" Mar 13 10:19:35 crc kubenswrapper[4930]: I0313 10:19:35.857144 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7181af2-26ac-48dd-988a-0b606b282234" containerName="registry-server" Mar 13 10:19:35 crc kubenswrapper[4930]: I0313 10:19:35.859581 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:35 crc kubenswrapper[4930]: I0313 10:19:35.875311 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4j8jm"] Mar 13 10:19:36 crc kubenswrapper[4930]: I0313 10:19:36.024354 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2wd2d\" (UniqueName: \"kubernetes.io/projected/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-kube-api-access-2wd2d\") pod \"certified-operators-4j8jm\" (UID: \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\") " pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:36 crc kubenswrapper[4930]: I0313 10:19:36.024514 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-catalog-content\") pod \"certified-operators-4j8jm\" (UID: \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\") " pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:36 crc kubenswrapper[4930]: I0313 10:19:36.024714 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-utilities\") pod \"certified-operators-4j8jm\" (UID: \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\") " pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:36 crc kubenswrapper[4930]: I0313 10:19:36.127564 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2wd2d\" (UniqueName: \"kubernetes.io/projected/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-kube-api-access-2wd2d\") pod \"certified-operators-4j8jm\" (UID: \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\") " pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:36 crc kubenswrapper[4930]: I0313 10:19:36.127809 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-catalog-content\") pod \"certified-operators-4j8jm\" (UID: \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\") " pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:36 crc kubenswrapper[4930]: I0313 10:19:36.128048 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-utilities\") pod \"certified-operators-4j8jm\" (UID: \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\") " pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:36 crc kubenswrapper[4930]: I0313 10:19:36.128542 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-catalog-content\") pod \"certified-operators-4j8jm\" (UID: \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\") " pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:36 crc kubenswrapper[4930]: I0313 10:19:36.128698 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-utilities\") pod \"certified-operators-4j8jm\" (UID: \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\") " pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:36 crc kubenswrapper[4930]: I0313 10:19:36.147047 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2wd2d\" (UniqueName: \"kubernetes.io/projected/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-kube-api-access-2wd2d\") pod \"certified-operators-4j8jm\" (UID: \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\") " pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:36 crc kubenswrapper[4930]: I0313 10:19:36.197829 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:36 crc kubenswrapper[4930]: I0313 10:19:36.837041 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4j8jm"] Mar 13 10:19:36 crc kubenswrapper[4930]: I0313 10:19:36.897825 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j8jm" event={"ID":"79eaba6a-3541-4e8d-9ef9-8a889ffbb398","Type":"ContainerStarted","Data":"1a7953e62c87fbb1601b02ad41b55c4a37937e90bbdc51a0db5fffd9aa20a104"} Mar 13 10:19:37 crc kubenswrapper[4930]: I0313 10:19:37.909336 4930 generic.go:334] "Generic (PLEG): container finished" podID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" containerID="41c1bc9cb0dc30cf6e766c6f0a230a00cb791e1287496caea7a8a8300f1b717d" exitCode=0 Mar 13 10:19:37 crc kubenswrapper[4930]: I0313 10:19:37.909462 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j8jm" event={"ID":"79eaba6a-3541-4e8d-9ef9-8a889ffbb398","Type":"ContainerDied","Data":"41c1bc9cb0dc30cf6e766c6f0a230a00cb791e1287496caea7a8a8300f1b717d"} Mar 13 10:19:37 crc kubenswrapper[4930]: I0313 10:19:37.911911 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:19:38 crc kubenswrapper[4930]: I0313 10:19:38.923910 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j8jm" event={"ID":"79eaba6a-3541-4e8d-9ef9-8a889ffbb398","Type":"ContainerStarted","Data":"2badeb4e6f1bad4331f08ac5249fc94c5a19b145f261f27b360a6309bb8ef6de"} Mar 13 10:19:41 crc kubenswrapper[4930]: I0313 10:19:41.959513 4930 generic.go:334] "Generic (PLEG): container finished" podID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" containerID="2badeb4e6f1bad4331f08ac5249fc94c5a19b145f261f27b360a6309bb8ef6de" exitCode=0 Mar 13 10:19:41 crc kubenswrapper[4930]: I0313 10:19:41.959612 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j8jm" event={"ID":"79eaba6a-3541-4e8d-9ef9-8a889ffbb398","Type":"ContainerDied","Data":"2badeb4e6f1bad4331f08ac5249fc94c5a19b145f261f27b360a6309bb8ef6de"} Mar 13 10:19:42 crc kubenswrapper[4930]: I0313 10:19:42.975258 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j8jm" event={"ID":"79eaba6a-3541-4e8d-9ef9-8a889ffbb398","Type":"ContainerStarted","Data":"b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f"} Mar 13 10:19:43 crc kubenswrapper[4930]: I0313 10:19:43.004206 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4j8jm" podStartSLOduration=3.390953069 podStartE2EDuration="8.004181688s" podCreationTimestamp="2026-03-13 10:19:35 +0000 UTC" firstStartedPulling="2026-03-13 10:19:37.911251645 +0000 UTC m=+4018.661166322" lastFinishedPulling="2026-03-13 10:19:42.524480254 +0000 UTC m=+4023.274394941" observedRunningTime="2026-03-13 10:19:42.994589867 +0000 UTC m=+4023.744504544" watchObservedRunningTime="2026-03-13 10:19:43.004181688 +0000 UTC m=+4023.754096365" Mar 13 10:19:46 crc kubenswrapper[4930]: I0313 10:19:46.198141 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:46 crc kubenswrapper[4930]: I0313 10:19:46.198712 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:47 crc kubenswrapper[4930]: I0313 10:19:47.248408 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-4j8jm" podUID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" containerName="registry-server" probeResult="failure" output=< Mar 13 10:19:47 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:19:47 crc kubenswrapper[4930]: > Mar 13 10:19:56 crc kubenswrapper[4930]: I0313 10:19:56.266115 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:56 crc kubenswrapper[4930]: I0313 10:19:56.325534 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:56 crc kubenswrapper[4930]: I0313 10:19:56.542750 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4j8jm"] Mar 13 10:19:58 crc kubenswrapper[4930]: I0313 10:19:58.250340 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4j8jm" podUID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" containerName="registry-server" containerID="cri-o://b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f" gracePeriod=2 Mar 13 10:19:58 crc kubenswrapper[4930]: I0313 10:19:58.818947 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:58 crc kubenswrapper[4930]: I0313 10:19:58.941311 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-catalog-content\") pod \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\" (UID: \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\") " Mar 13 10:19:58 crc kubenswrapper[4930]: I0313 10:19:58.941361 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2wd2d\" (UniqueName: \"kubernetes.io/projected/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-kube-api-access-2wd2d\") pod \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\" (UID: \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\") " Mar 13 10:19:58 crc kubenswrapper[4930]: I0313 10:19:58.941525 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-utilities\") pod \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\" (UID: \"79eaba6a-3541-4e8d-9ef9-8a889ffbb398\") " Mar 13 10:19:58 crc kubenswrapper[4930]: I0313 10:19:58.942930 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-utilities" (OuterVolumeSpecName: "utilities") pod "79eaba6a-3541-4e8d-9ef9-8a889ffbb398" (UID: "79eaba6a-3541-4e8d-9ef9-8a889ffbb398"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:19:58 crc kubenswrapper[4930]: I0313 10:19:58.951277 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-kube-api-access-2wd2d" (OuterVolumeSpecName: "kube-api-access-2wd2d") pod "79eaba6a-3541-4e8d-9ef9-8a889ffbb398" (UID: "79eaba6a-3541-4e8d-9ef9-8a889ffbb398"). InnerVolumeSpecName "kube-api-access-2wd2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.013758 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "79eaba6a-3541-4e8d-9ef9-8a889ffbb398" (UID: "79eaba6a-3541-4e8d-9ef9-8a889ffbb398"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.044599 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.044636 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2wd2d\" (UniqueName: \"kubernetes.io/projected/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-kube-api-access-2wd2d\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.044646 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/79eaba6a-3541-4e8d-9ef9-8a889ffbb398-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.265826 4930 generic.go:334] "Generic (PLEG): container finished" podID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" containerID="b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f" exitCode=0 Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.265876 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j8jm" event={"ID":"79eaba6a-3541-4e8d-9ef9-8a889ffbb398","Type":"ContainerDied","Data":"b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f"} Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.265945 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4j8jm" event={"ID":"79eaba6a-3541-4e8d-9ef9-8a889ffbb398","Type":"ContainerDied","Data":"1a7953e62c87fbb1601b02ad41b55c4a37937e90bbdc51a0db5fffd9aa20a104"} Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.265962 4930 scope.go:117] "RemoveContainer" containerID="b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.267308 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4j8jm" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.292233 4930 scope.go:117] "RemoveContainer" containerID="2badeb4e6f1bad4331f08ac5249fc94c5a19b145f261f27b360a6309bb8ef6de" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.316996 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4j8jm"] Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.331343 4930 scope.go:117] "RemoveContainer" containerID="41c1bc9cb0dc30cf6e766c6f0a230a00cb791e1287496caea7a8a8300f1b717d" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.335821 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4j8jm"] Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.408996 4930 scope.go:117] "RemoveContainer" containerID="b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f" Mar 13 10:19:59 crc kubenswrapper[4930]: E0313 10:19:59.411517 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f\": container with ID starting with b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f not found: ID does not exist" containerID="b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.411564 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f"} err="failed to get container status \"b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f\": rpc error: code = NotFound desc = could not find container \"b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f\": container with ID starting with b8d9bd3df1577499e3afe545446cf166e97016b6a1fa2e53ec1fb7f4b420392f not found: ID does not exist" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.411589 4930 scope.go:117] "RemoveContainer" containerID="2badeb4e6f1bad4331f08ac5249fc94c5a19b145f261f27b360a6309bb8ef6de" Mar 13 10:19:59 crc kubenswrapper[4930]: E0313 10:19:59.413202 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2badeb4e6f1bad4331f08ac5249fc94c5a19b145f261f27b360a6309bb8ef6de\": container with ID starting with 2badeb4e6f1bad4331f08ac5249fc94c5a19b145f261f27b360a6309bb8ef6de not found: ID does not exist" containerID="2badeb4e6f1bad4331f08ac5249fc94c5a19b145f261f27b360a6309bb8ef6de" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.413238 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2badeb4e6f1bad4331f08ac5249fc94c5a19b145f261f27b360a6309bb8ef6de"} err="failed to get container status \"2badeb4e6f1bad4331f08ac5249fc94c5a19b145f261f27b360a6309bb8ef6de\": rpc error: code = NotFound desc = could not find container \"2badeb4e6f1bad4331f08ac5249fc94c5a19b145f261f27b360a6309bb8ef6de\": container with ID starting with 2badeb4e6f1bad4331f08ac5249fc94c5a19b145f261f27b360a6309bb8ef6de not found: ID does not exist" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.413263 4930 scope.go:117] "RemoveContainer" containerID="41c1bc9cb0dc30cf6e766c6f0a230a00cb791e1287496caea7a8a8300f1b717d" Mar 13 10:19:59 crc kubenswrapper[4930]: E0313 10:19:59.413749 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41c1bc9cb0dc30cf6e766c6f0a230a00cb791e1287496caea7a8a8300f1b717d\": container with ID starting with 41c1bc9cb0dc30cf6e766c6f0a230a00cb791e1287496caea7a8a8300f1b717d not found: ID does not exist" containerID="41c1bc9cb0dc30cf6e766c6f0a230a00cb791e1287496caea7a8a8300f1b717d" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.413773 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41c1bc9cb0dc30cf6e766c6f0a230a00cb791e1287496caea7a8a8300f1b717d"} err="failed to get container status \"41c1bc9cb0dc30cf6e766c6f0a230a00cb791e1287496caea7a8a8300f1b717d\": rpc error: code = NotFound desc = could not find container \"41c1bc9cb0dc30cf6e766c6f0a230a00cb791e1287496caea7a8a8300f1b717d\": container with ID starting with 41c1bc9cb0dc30cf6e766c6f0a230a00cb791e1287496caea7a8a8300f1b717d not found: ID does not exist" Mar 13 10:19:59 crc kubenswrapper[4930]: I0313 10:19:59.986285 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" path="/var/lib/kubelet/pods/79eaba6a-3541-4e8d-9ef9-8a889ffbb398/volumes" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.145448 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556620-7zrwd"] Mar 13 10:20:00 crc kubenswrapper[4930]: E0313 10:20:00.145978 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" containerName="extract-utilities" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.145997 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" containerName="extract-utilities" Mar 13 10:20:00 crc kubenswrapper[4930]: E0313 10:20:00.146052 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" containerName="extract-content" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.146061 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" containerName="extract-content" Mar 13 10:20:00 crc kubenswrapper[4930]: E0313 10:20:00.146107 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" containerName="registry-server" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.146115 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" containerName="registry-server" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.146322 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="79eaba6a-3541-4e8d-9ef9-8a889ffbb398" containerName="registry-server" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.147158 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556620-7zrwd" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.150868 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.150895 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.150955 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.158252 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556620-7zrwd"] Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.171857 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnwrc\" (UniqueName: \"kubernetes.io/projected/0715803e-233a-45bd-9549-a619e4767274-kube-api-access-qnwrc\") pod \"auto-csr-approver-29556620-7zrwd\" (UID: \"0715803e-233a-45bd-9549-a619e4767274\") " pod="openshift-infra/auto-csr-approver-29556620-7zrwd" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.274733 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnwrc\" (UniqueName: \"kubernetes.io/projected/0715803e-233a-45bd-9549-a619e4767274-kube-api-access-qnwrc\") pod \"auto-csr-approver-29556620-7zrwd\" (UID: \"0715803e-233a-45bd-9549-a619e4767274\") " pod="openshift-infra/auto-csr-approver-29556620-7zrwd" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.319240 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnwrc\" (UniqueName: \"kubernetes.io/projected/0715803e-233a-45bd-9549-a619e4767274-kube-api-access-qnwrc\") pod \"auto-csr-approver-29556620-7zrwd\" (UID: \"0715803e-233a-45bd-9549-a619e4767274\") " pod="openshift-infra/auto-csr-approver-29556620-7zrwd" Mar 13 10:20:00 crc kubenswrapper[4930]: I0313 10:20:00.468885 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556620-7zrwd" Mar 13 10:20:01 crc kubenswrapper[4930]: I0313 10:20:01.013692 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556620-7zrwd"] Mar 13 10:20:01 crc kubenswrapper[4930]: I0313 10:20:01.289082 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556620-7zrwd" event={"ID":"0715803e-233a-45bd-9549-a619e4767274","Type":"ContainerStarted","Data":"b4c0a296a467421222c207fbfb3449036b76ad30cc81cb69f43908d158912769"} Mar 13 10:20:03 crc kubenswrapper[4930]: I0313 10:20:03.310296 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556620-7zrwd" event={"ID":"0715803e-233a-45bd-9549-a619e4767274","Type":"ContainerStarted","Data":"cdad3e9825ddafe8ba60cd1b7bfe35984314001bafddabd324fa132df2e34540"} Mar 13 10:20:03 crc kubenswrapper[4930]: I0313 10:20:03.337381 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556620-7zrwd" podStartSLOduration=1.8409672879999999 podStartE2EDuration="3.337360651s" podCreationTimestamp="2026-03-13 10:20:00 +0000 UTC" firstStartedPulling="2026-03-13 10:20:01.005613119 +0000 UTC m=+4041.755527796" lastFinishedPulling="2026-03-13 10:20:02.502006472 +0000 UTC m=+4043.251921159" observedRunningTime="2026-03-13 10:20:03.324052657 +0000 UTC m=+4044.073967334" watchObservedRunningTime="2026-03-13 10:20:03.337360651 +0000 UTC m=+4044.087275328" Mar 13 10:20:04 crc kubenswrapper[4930]: I0313 10:20:04.326395 4930 generic.go:334] "Generic (PLEG): container finished" podID="0715803e-233a-45bd-9549-a619e4767274" containerID="cdad3e9825ddafe8ba60cd1b7bfe35984314001bafddabd324fa132df2e34540" exitCode=0 Mar 13 10:20:04 crc kubenswrapper[4930]: I0313 10:20:04.326474 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556620-7zrwd" event={"ID":"0715803e-233a-45bd-9549-a619e4767274","Type":"ContainerDied","Data":"cdad3e9825ddafe8ba60cd1b7bfe35984314001bafddabd324fa132df2e34540"} Mar 13 10:20:05 crc kubenswrapper[4930]: I0313 10:20:05.863221 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556620-7zrwd" Mar 13 10:20:06 crc kubenswrapper[4930]: I0313 10:20:06.019739 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnwrc\" (UniqueName: \"kubernetes.io/projected/0715803e-233a-45bd-9549-a619e4767274-kube-api-access-qnwrc\") pod \"0715803e-233a-45bd-9549-a619e4767274\" (UID: \"0715803e-233a-45bd-9549-a619e4767274\") " Mar 13 10:20:06 crc kubenswrapper[4930]: I0313 10:20:06.025695 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0715803e-233a-45bd-9549-a619e4767274-kube-api-access-qnwrc" (OuterVolumeSpecName: "kube-api-access-qnwrc") pod "0715803e-233a-45bd-9549-a619e4767274" (UID: "0715803e-233a-45bd-9549-a619e4767274"). InnerVolumeSpecName "kube-api-access-qnwrc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:20:06 crc kubenswrapper[4930]: I0313 10:20:06.126925 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnwrc\" (UniqueName: \"kubernetes.io/projected/0715803e-233a-45bd-9549-a619e4767274-kube-api-access-qnwrc\") on node \"crc\" DevicePath \"\"" Mar 13 10:20:06 crc kubenswrapper[4930]: I0313 10:20:06.350885 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556620-7zrwd" event={"ID":"0715803e-233a-45bd-9549-a619e4767274","Type":"ContainerDied","Data":"b4c0a296a467421222c207fbfb3449036b76ad30cc81cb69f43908d158912769"} Mar 13 10:20:06 crc kubenswrapper[4930]: I0313 10:20:06.351168 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4c0a296a467421222c207fbfb3449036b76ad30cc81cb69f43908d158912769" Mar 13 10:20:06 crc kubenswrapper[4930]: I0313 10:20:06.350982 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556620-7zrwd" Mar 13 10:20:06 crc kubenswrapper[4930]: I0313 10:20:06.409909 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556614-rc9f4"] Mar 13 10:20:06 crc kubenswrapper[4930]: I0313 10:20:06.422206 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556614-rc9f4"] Mar 13 10:20:07 crc kubenswrapper[4930]: I0313 10:20:07.984371 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4b1588b-1d42-41c4-ab24-35ae3edea14f" path="/var/lib/kubelet/pods/c4b1588b-1d42-41c4-ab24-35ae3edea14f/volumes" Mar 13 10:20:32 crc kubenswrapper[4930]: I0313 10:20:32.989399 4930 scope.go:117] "RemoveContainer" containerID="8c3b57e581f809ab8fdbf2a8a1c5e433242a002245cbf8aaf34b54b70566cfc4" Mar 13 10:20:42 crc kubenswrapper[4930]: I0313 10:20:42.308819 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:20:42 crc kubenswrapper[4930]: I0313 10:20:42.309371 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:21:12 crc kubenswrapper[4930]: I0313 10:21:12.308391 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:21:12 crc kubenswrapper[4930]: I0313 10:21:12.309616 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:21:42 crc kubenswrapper[4930]: I0313 10:21:42.308325 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:21:42 crc kubenswrapper[4930]: I0313 10:21:42.308884 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:21:42 crc kubenswrapper[4930]: I0313 10:21:42.308929 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 10:21:42 crc kubenswrapper[4930]: I0313 10:21:42.309791 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a2dd0e422956ad3107b66f8a5d96648c6e90c0a7f1dd96ff6dcb9a41253b535b"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:21:42 crc kubenswrapper[4930]: I0313 10:21:42.309848 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://a2dd0e422956ad3107b66f8a5d96648c6e90c0a7f1dd96ff6dcb9a41253b535b" gracePeriod=600 Mar 13 10:21:43 crc kubenswrapper[4930]: I0313 10:21:43.415266 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="a2dd0e422956ad3107b66f8a5d96648c6e90c0a7f1dd96ff6dcb9a41253b535b" exitCode=0 Mar 13 10:21:43 crc kubenswrapper[4930]: I0313 10:21:43.415713 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"a2dd0e422956ad3107b66f8a5d96648c6e90c0a7f1dd96ff6dcb9a41253b535b"} Mar 13 10:21:43 crc kubenswrapper[4930]: I0313 10:21:43.415741 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db"} Mar 13 10:21:43 crc kubenswrapper[4930]: I0313 10:21:43.415756 4930 scope.go:117] "RemoveContainer" containerID="02d9fc64c8364a74d2ec2b50e8954f6bf7460074f4e402cbe9d7a54ec30db5cf" Mar 13 10:22:00 crc kubenswrapper[4930]: I0313 10:22:00.149878 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556622-fn42m"] Mar 13 10:22:00 crc kubenswrapper[4930]: E0313 10:22:00.150990 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0715803e-233a-45bd-9549-a619e4767274" containerName="oc" Mar 13 10:22:00 crc kubenswrapper[4930]: I0313 10:22:00.151005 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="0715803e-233a-45bd-9549-a619e4767274" containerName="oc" Mar 13 10:22:00 crc kubenswrapper[4930]: I0313 10:22:00.151234 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="0715803e-233a-45bd-9549-a619e4767274" containerName="oc" Mar 13 10:22:00 crc kubenswrapper[4930]: I0313 10:22:00.152243 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556622-fn42m" Mar 13 10:22:00 crc kubenswrapper[4930]: I0313 10:22:00.155760 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:22:00 crc kubenswrapper[4930]: I0313 10:22:00.156337 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:22:00 crc kubenswrapper[4930]: I0313 10:22:00.157611 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:22:00 crc kubenswrapper[4930]: I0313 10:22:00.166326 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556622-fn42m"] Mar 13 10:22:00 crc kubenswrapper[4930]: I0313 10:22:00.263108 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8spvm\" (UniqueName: \"kubernetes.io/projected/fca6144f-3b4c-4d42-af91-3bcb2a4fd06d-kube-api-access-8spvm\") pod \"auto-csr-approver-29556622-fn42m\" (UID: \"fca6144f-3b4c-4d42-af91-3bcb2a4fd06d\") " pod="openshift-infra/auto-csr-approver-29556622-fn42m" Mar 13 10:22:00 crc kubenswrapper[4930]: I0313 10:22:00.365812 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8spvm\" (UniqueName: \"kubernetes.io/projected/fca6144f-3b4c-4d42-af91-3bcb2a4fd06d-kube-api-access-8spvm\") pod \"auto-csr-approver-29556622-fn42m\" (UID: \"fca6144f-3b4c-4d42-af91-3bcb2a4fd06d\") " pod="openshift-infra/auto-csr-approver-29556622-fn42m" Mar 13 10:22:00 crc kubenswrapper[4930]: I0313 10:22:00.400680 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8spvm\" (UniqueName: \"kubernetes.io/projected/fca6144f-3b4c-4d42-af91-3bcb2a4fd06d-kube-api-access-8spvm\") pod \"auto-csr-approver-29556622-fn42m\" (UID: \"fca6144f-3b4c-4d42-af91-3bcb2a4fd06d\") " pod="openshift-infra/auto-csr-approver-29556622-fn42m" Mar 13 10:22:00 crc kubenswrapper[4930]: I0313 10:22:00.475137 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556622-fn42m" Mar 13 10:22:01 crc kubenswrapper[4930]: I0313 10:22:01.600535 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556622-fn42m"] Mar 13 10:22:01 crc kubenswrapper[4930]: I0313 10:22:01.620457 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556622-fn42m" event={"ID":"fca6144f-3b4c-4d42-af91-3bcb2a4fd06d","Type":"ContainerStarted","Data":"a4cddfa596bfd3a2ef8f7aad7d4a378801d191926acc40fa614b84db0f27fcf3"} Mar 13 10:22:03 crc kubenswrapper[4930]: I0313 10:22:03.658794 4930 generic.go:334] "Generic (PLEG): container finished" podID="fca6144f-3b4c-4d42-af91-3bcb2a4fd06d" containerID="3df03912507e46221ac936ddf7139e3c087745163dcd99663ea52953b0a86ea2" exitCode=0 Mar 13 10:22:03 crc kubenswrapper[4930]: I0313 10:22:03.659457 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556622-fn42m" event={"ID":"fca6144f-3b4c-4d42-af91-3bcb2a4fd06d","Type":"ContainerDied","Data":"3df03912507e46221ac936ddf7139e3c087745163dcd99663ea52953b0a86ea2"} Mar 13 10:22:05 crc kubenswrapper[4930]: I0313 10:22:05.177553 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556622-fn42m" Mar 13 10:22:05 crc kubenswrapper[4930]: I0313 10:22:05.211382 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8spvm\" (UniqueName: \"kubernetes.io/projected/fca6144f-3b4c-4d42-af91-3bcb2a4fd06d-kube-api-access-8spvm\") pod \"fca6144f-3b4c-4d42-af91-3bcb2a4fd06d\" (UID: \"fca6144f-3b4c-4d42-af91-3bcb2a4fd06d\") " Mar 13 10:22:05 crc kubenswrapper[4930]: I0313 10:22:05.221045 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fca6144f-3b4c-4d42-af91-3bcb2a4fd06d-kube-api-access-8spvm" (OuterVolumeSpecName: "kube-api-access-8spvm") pod "fca6144f-3b4c-4d42-af91-3bcb2a4fd06d" (UID: "fca6144f-3b4c-4d42-af91-3bcb2a4fd06d"). InnerVolumeSpecName "kube-api-access-8spvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:22:05 crc kubenswrapper[4930]: I0313 10:22:05.317275 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8spvm\" (UniqueName: \"kubernetes.io/projected/fca6144f-3b4c-4d42-af91-3bcb2a4fd06d-kube-api-access-8spvm\") on node \"crc\" DevicePath \"\"" Mar 13 10:22:05 crc kubenswrapper[4930]: I0313 10:22:05.681651 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556622-fn42m" event={"ID":"fca6144f-3b4c-4d42-af91-3bcb2a4fd06d","Type":"ContainerDied","Data":"a4cddfa596bfd3a2ef8f7aad7d4a378801d191926acc40fa614b84db0f27fcf3"} Mar 13 10:22:05 crc kubenswrapper[4930]: I0313 10:22:05.681702 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556622-fn42m" Mar 13 10:22:05 crc kubenswrapper[4930]: I0313 10:22:05.681720 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4cddfa596bfd3a2ef8f7aad7d4a378801d191926acc40fa614b84db0f27fcf3" Mar 13 10:22:06 crc kubenswrapper[4930]: I0313 10:22:06.254782 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556616-qfhn8"] Mar 13 10:22:06 crc kubenswrapper[4930]: I0313 10:22:06.266902 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556616-qfhn8"] Mar 13 10:22:07 crc kubenswrapper[4930]: I0313 10:22:07.983798 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf65e4e5-70cc-4d4d-a31e-010272edcf0f" path="/var/lib/kubelet/pods/bf65e4e5-70cc-4d4d-a31e-010272edcf0f/volumes" Mar 13 10:22:33 crc kubenswrapper[4930]: I0313 10:22:33.123198 4930 scope.go:117] "RemoveContainer" containerID="b516251ebbf3124180be85cc79ac40c339769a56f1d99899996843200996dec8" Mar 13 10:23:42 crc kubenswrapper[4930]: I0313 10:23:42.308683 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:23:42 crc kubenswrapper[4930]: I0313 10:23:42.309105 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.150649 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556624-9rqsx"] Mar 13 10:24:00 crc kubenswrapper[4930]: E0313 10:24:00.152843 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fca6144f-3b4c-4d42-af91-3bcb2a4fd06d" containerName="oc" Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.152926 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="fca6144f-3b4c-4d42-af91-3bcb2a4fd06d" containerName="oc" Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.153217 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="fca6144f-3b4c-4d42-af91-3bcb2a4fd06d" containerName="oc" Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.154277 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556624-9rqsx" Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.161024 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.161096 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.161124 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.163393 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556624-9rqsx"] Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.213949 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzpj7\" (UniqueName: \"kubernetes.io/projected/ffa5b2b3-62be-4df6-a444-a1634dc993cb-kube-api-access-fzpj7\") pod \"auto-csr-approver-29556624-9rqsx\" (UID: \"ffa5b2b3-62be-4df6-a444-a1634dc993cb\") " pod="openshift-infra/auto-csr-approver-29556624-9rqsx" Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.316092 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzpj7\" (UniqueName: \"kubernetes.io/projected/ffa5b2b3-62be-4df6-a444-a1634dc993cb-kube-api-access-fzpj7\") pod \"auto-csr-approver-29556624-9rqsx\" (UID: \"ffa5b2b3-62be-4df6-a444-a1634dc993cb\") " pod="openshift-infra/auto-csr-approver-29556624-9rqsx" Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.343038 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzpj7\" (UniqueName: \"kubernetes.io/projected/ffa5b2b3-62be-4df6-a444-a1634dc993cb-kube-api-access-fzpj7\") pod \"auto-csr-approver-29556624-9rqsx\" (UID: \"ffa5b2b3-62be-4df6-a444-a1634dc993cb\") " pod="openshift-infra/auto-csr-approver-29556624-9rqsx" Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.492173 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556624-9rqsx" Mar 13 10:24:00 crc kubenswrapper[4930]: I0313 10:24:00.966494 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556624-9rqsx"] Mar 13 10:24:01 crc kubenswrapper[4930]: I0313 10:24:01.986705 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556624-9rqsx" event={"ID":"ffa5b2b3-62be-4df6-a444-a1634dc993cb","Type":"ContainerStarted","Data":"a8245835af6a7c317922db865c272dc0275075377ad136943e2623ea3a935a4a"} Mar 13 10:24:02 crc kubenswrapper[4930]: I0313 10:24:02.994275 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556624-9rqsx" event={"ID":"ffa5b2b3-62be-4df6-a444-a1634dc993cb","Type":"ContainerStarted","Data":"7d6331a452b3dcbfaeaf0f6a4d7d4db4a044511b9751df063cffc1996d1d433c"} Mar 13 10:24:03 crc kubenswrapper[4930]: I0313 10:24:03.025561 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556624-9rqsx" podStartSLOduration=1.9476893309999999 podStartE2EDuration="3.025537283s" podCreationTimestamp="2026-03-13 10:24:00 +0000 UTC" firstStartedPulling="2026-03-13 10:24:01.470022565 +0000 UTC m=+4282.219937242" lastFinishedPulling="2026-03-13 10:24:02.547870517 +0000 UTC m=+4283.297785194" observedRunningTime="2026-03-13 10:24:03.0095496 +0000 UTC m=+4283.759464277" watchObservedRunningTime="2026-03-13 10:24:03.025537283 +0000 UTC m=+4283.775451960" Mar 13 10:24:04 crc kubenswrapper[4930]: I0313 10:24:04.006840 4930 generic.go:334] "Generic (PLEG): container finished" podID="ffa5b2b3-62be-4df6-a444-a1634dc993cb" containerID="7d6331a452b3dcbfaeaf0f6a4d7d4db4a044511b9751df063cffc1996d1d433c" exitCode=0 Mar 13 10:24:04 crc kubenswrapper[4930]: I0313 10:24:04.006881 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556624-9rqsx" event={"ID":"ffa5b2b3-62be-4df6-a444-a1634dc993cb","Type":"ContainerDied","Data":"7d6331a452b3dcbfaeaf0f6a4d7d4db4a044511b9751df063cffc1996d1d433c"} Mar 13 10:24:05 crc kubenswrapper[4930]: I0313 10:24:05.450902 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556624-9rqsx" Mar 13 10:24:05 crc kubenswrapper[4930]: I0313 10:24:05.552194 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzpj7\" (UniqueName: \"kubernetes.io/projected/ffa5b2b3-62be-4df6-a444-a1634dc993cb-kube-api-access-fzpj7\") pod \"ffa5b2b3-62be-4df6-a444-a1634dc993cb\" (UID: \"ffa5b2b3-62be-4df6-a444-a1634dc993cb\") " Mar 13 10:24:05 crc kubenswrapper[4930]: I0313 10:24:05.562060 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ffa5b2b3-62be-4df6-a444-a1634dc993cb-kube-api-access-fzpj7" (OuterVolumeSpecName: "kube-api-access-fzpj7") pod "ffa5b2b3-62be-4df6-a444-a1634dc993cb" (UID: "ffa5b2b3-62be-4df6-a444-a1634dc993cb"). InnerVolumeSpecName "kube-api-access-fzpj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:24:05 crc kubenswrapper[4930]: I0313 10:24:05.654986 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzpj7\" (UniqueName: \"kubernetes.io/projected/ffa5b2b3-62be-4df6-a444-a1634dc993cb-kube-api-access-fzpj7\") on node \"crc\" DevicePath \"\"" Mar 13 10:24:06 crc kubenswrapper[4930]: I0313 10:24:06.032116 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556624-9rqsx" event={"ID":"ffa5b2b3-62be-4df6-a444-a1634dc993cb","Type":"ContainerDied","Data":"a8245835af6a7c317922db865c272dc0275075377ad136943e2623ea3a935a4a"} Mar 13 10:24:06 crc kubenswrapper[4930]: I0313 10:24:06.032173 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8245835af6a7c317922db865c272dc0275075377ad136943e2623ea3a935a4a" Mar 13 10:24:06 crc kubenswrapper[4930]: I0313 10:24:06.032283 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556624-9rqsx" Mar 13 10:24:06 crc kubenswrapper[4930]: I0313 10:24:06.087837 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556618-bh7lw"] Mar 13 10:24:06 crc kubenswrapper[4930]: I0313 10:24:06.100801 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556618-bh7lw"] Mar 13 10:24:07 crc kubenswrapper[4930]: I0313 10:24:07.985422 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1" path="/var/lib/kubelet/pods/5f0c052e-0114-4aa3-bb68-fb5ebe2b01b1/volumes" Mar 13 10:24:12 crc kubenswrapper[4930]: I0313 10:24:12.308690 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:24:12 crc kubenswrapper[4930]: I0313 10:24:12.309137 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:24:33 crc kubenswrapper[4930]: I0313 10:24:33.852902 4930 scope.go:117] "RemoveContainer" containerID="b1e995e5f882c69d87520b16aba7b659e90c0b3e164b93342f09038af30ac081" Mar 13 10:24:42 crc kubenswrapper[4930]: I0313 10:24:42.308409 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:24:42 crc kubenswrapper[4930]: I0313 10:24:42.308940 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:24:42 crc kubenswrapper[4930]: I0313 10:24:42.308996 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 10:24:42 crc kubenswrapper[4930]: I0313 10:24:42.309966 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:24:42 crc kubenswrapper[4930]: I0313 10:24:42.310035 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" gracePeriod=600 Mar 13 10:24:42 crc kubenswrapper[4930]: E0313 10:24:42.436087 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:24:43 crc kubenswrapper[4930]: I0313 10:24:43.441275 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" exitCode=0 Mar 13 10:24:43 crc kubenswrapper[4930]: I0313 10:24:43.441346 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db"} Mar 13 10:24:43 crc kubenswrapper[4930]: I0313 10:24:43.441603 4930 scope.go:117] "RemoveContainer" containerID="a2dd0e422956ad3107b66f8a5d96648c6e90c0a7f1dd96ff6dcb9a41253b535b" Mar 13 10:24:43 crc kubenswrapper[4930]: I0313 10:24:43.442299 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:24:43 crc kubenswrapper[4930]: E0313 10:24:43.442609 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:24:57 crc kubenswrapper[4930]: I0313 10:24:57.972331 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:24:57 crc kubenswrapper[4930]: E0313 10:24:57.973503 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:25:08 crc kubenswrapper[4930]: I0313 10:25:08.972493 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:25:08 crc kubenswrapper[4930]: E0313 10:25:08.973928 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:25:23 crc kubenswrapper[4930]: I0313 10:25:23.971297 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:25:23 crc kubenswrapper[4930]: E0313 10:25:23.972123 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:25:38 crc kubenswrapper[4930]: I0313 10:25:38.971753 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:25:38 crc kubenswrapper[4930]: E0313 10:25:38.972989 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:25:50 crc kubenswrapper[4930]: I0313 10:25:50.970902 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:25:50 crc kubenswrapper[4930]: E0313 10:25:50.971728 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.115000 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zzmpk"] Mar 13 10:25:53 crc kubenswrapper[4930]: E0313 10:25:53.116947 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ffa5b2b3-62be-4df6-a444-a1634dc993cb" containerName="oc" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.116973 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="ffa5b2b3-62be-4df6-a444-a1634dc993cb" containerName="oc" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.118264 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="ffa5b2b3-62be-4df6-a444-a1634dc993cb" containerName="oc" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.137834 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.187827 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zzmpk"] Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.326277 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd737da-1d04-430b-8ee4-573fd5da59a3-catalog-content\") pod \"redhat-operators-zzmpk\" (UID: \"3cd737da-1d04-430b-8ee4-573fd5da59a3\") " pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.326365 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd737da-1d04-430b-8ee4-573fd5da59a3-utilities\") pod \"redhat-operators-zzmpk\" (UID: \"3cd737da-1d04-430b-8ee4-573fd5da59a3\") " pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.326422 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kxbm\" (UniqueName: \"kubernetes.io/projected/3cd737da-1d04-430b-8ee4-573fd5da59a3-kube-api-access-6kxbm\") pod \"redhat-operators-zzmpk\" (UID: \"3cd737da-1d04-430b-8ee4-573fd5da59a3\") " pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.429048 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd737da-1d04-430b-8ee4-573fd5da59a3-catalog-content\") pod \"redhat-operators-zzmpk\" (UID: \"3cd737da-1d04-430b-8ee4-573fd5da59a3\") " pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.429127 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd737da-1d04-430b-8ee4-573fd5da59a3-utilities\") pod \"redhat-operators-zzmpk\" (UID: \"3cd737da-1d04-430b-8ee4-573fd5da59a3\") " pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.429200 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kxbm\" (UniqueName: \"kubernetes.io/projected/3cd737da-1d04-430b-8ee4-573fd5da59a3-kube-api-access-6kxbm\") pod \"redhat-operators-zzmpk\" (UID: \"3cd737da-1d04-430b-8ee4-573fd5da59a3\") " pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.429784 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd737da-1d04-430b-8ee4-573fd5da59a3-catalog-content\") pod \"redhat-operators-zzmpk\" (UID: \"3cd737da-1d04-430b-8ee4-573fd5da59a3\") " pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.429823 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd737da-1d04-430b-8ee4-573fd5da59a3-utilities\") pod \"redhat-operators-zzmpk\" (UID: \"3cd737da-1d04-430b-8ee4-573fd5da59a3\") " pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.765702 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kxbm\" (UniqueName: \"kubernetes.io/projected/3cd737da-1d04-430b-8ee4-573fd5da59a3-kube-api-access-6kxbm\") pod \"redhat-operators-zzmpk\" (UID: \"3cd737da-1d04-430b-8ee4-573fd5da59a3\") " pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:25:53 crc kubenswrapper[4930]: I0313 10:25:53.773857 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:25:54 crc kubenswrapper[4930]: I0313 10:25:54.426933 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zzmpk"] Mar 13 10:25:55 crc kubenswrapper[4930]: I0313 10:25:55.197596 4930 generic.go:334] "Generic (PLEG): container finished" podID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerID="167c36ac1141c75b6dd3f36900d6310aadf25272286e32c4e8aeacf1396169d6" exitCode=0 Mar 13 10:25:55 crc kubenswrapper[4930]: I0313 10:25:55.197655 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzmpk" event={"ID":"3cd737da-1d04-430b-8ee4-573fd5da59a3","Type":"ContainerDied","Data":"167c36ac1141c75b6dd3f36900d6310aadf25272286e32c4e8aeacf1396169d6"} Mar 13 10:25:55 crc kubenswrapper[4930]: I0313 10:25:55.198190 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzmpk" event={"ID":"3cd737da-1d04-430b-8ee4-573fd5da59a3","Type":"ContainerStarted","Data":"bb5b5e94ec31e0e4e441c91575d6001bc31cb457dc2988406b5e8a717363a869"} Mar 13 10:25:55 crc kubenswrapper[4930]: I0313 10:25:55.200222 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:25:56 crc kubenswrapper[4930]: I0313 10:25:56.210329 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzmpk" event={"ID":"3cd737da-1d04-430b-8ee4-573fd5da59a3","Type":"ContainerStarted","Data":"b8cd4ecf098534fe7483d6c7617bb3a4258608ca7b6d528cf09606a735331458"} Mar 13 10:26:00 crc kubenswrapper[4930]: I0313 10:26:00.146191 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556626-bfbhx"] Mar 13 10:26:00 crc kubenswrapper[4930]: I0313 10:26:00.148698 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556626-bfbhx" Mar 13 10:26:00 crc kubenswrapper[4930]: I0313 10:26:00.154156 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:26:00 crc kubenswrapper[4930]: I0313 10:26:00.157842 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:26:00 crc kubenswrapper[4930]: I0313 10:26:00.158019 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:26:00 crc kubenswrapper[4930]: I0313 10:26:00.158179 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556626-bfbhx"] Mar 13 10:26:00 crc kubenswrapper[4930]: I0313 10:26:00.246685 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55w4n\" (UniqueName: \"kubernetes.io/projected/44ba6b14-c48d-443a-bb78-cf3e560f3a46-kube-api-access-55w4n\") pod \"auto-csr-approver-29556626-bfbhx\" (UID: \"44ba6b14-c48d-443a-bb78-cf3e560f3a46\") " pod="openshift-infra/auto-csr-approver-29556626-bfbhx" Mar 13 10:26:00 crc kubenswrapper[4930]: I0313 10:26:00.349514 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55w4n\" (UniqueName: \"kubernetes.io/projected/44ba6b14-c48d-443a-bb78-cf3e560f3a46-kube-api-access-55w4n\") pod \"auto-csr-approver-29556626-bfbhx\" (UID: \"44ba6b14-c48d-443a-bb78-cf3e560f3a46\") " pod="openshift-infra/auto-csr-approver-29556626-bfbhx" Mar 13 10:26:00 crc kubenswrapper[4930]: I0313 10:26:00.418331 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55w4n\" (UniqueName: \"kubernetes.io/projected/44ba6b14-c48d-443a-bb78-cf3e560f3a46-kube-api-access-55w4n\") pod \"auto-csr-approver-29556626-bfbhx\" (UID: \"44ba6b14-c48d-443a-bb78-cf3e560f3a46\") " pod="openshift-infra/auto-csr-approver-29556626-bfbhx" Mar 13 10:26:00 crc kubenswrapper[4930]: I0313 10:26:00.469287 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556626-bfbhx" Mar 13 10:26:01 crc kubenswrapper[4930]: I0313 10:26:01.034631 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556626-bfbhx"] Mar 13 10:26:01 crc kubenswrapper[4930]: I0313 10:26:01.263711 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556626-bfbhx" event={"ID":"44ba6b14-c48d-443a-bb78-cf3e560f3a46","Type":"ContainerStarted","Data":"59915d0e4c274c8d763699a4feb3cdaae8868caa77291c92f4815b3b832242ed"} Mar 13 10:26:02 crc kubenswrapper[4930]: I0313 10:26:02.281239 4930 generic.go:334] "Generic (PLEG): container finished" podID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerID="b8cd4ecf098534fe7483d6c7617bb3a4258608ca7b6d528cf09606a735331458" exitCode=0 Mar 13 10:26:02 crc kubenswrapper[4930]: I0313 10:26:02.281296 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzmpk" event={"ID":"3cd737da-1d04-430b-8ee4-573fd5da59a3","Type":"ContainerDied","Data":"b8cd4ecf098534fe7483d6c7617bb3a4258608ca7b6d528cf09606a735331458"} Mar 13 10:26:03 crc kubenswrapper[4930]: I0313 10:26:03.294265 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzmpk" event={"ID":"3cd737da-1d04-430b-8ee4-573fd5da59a3","Type":"ContainerStarted","Data":"da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01"} Mar 13 10:26:03 crc kubenswrapper[4930]: I0313 10:26:03.296083 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556626-bfbhx" event={"ID":"44ba6b14-c48d-443a-bb78-cf3e560f3a46","Type":"ContainerStarted","Data":"c62ea15f39c602bfe3561c09b828e926f13747cd5106b1bae48468bf120409a3"} Mar 13 10:26:03 crc kubenswrapper[4930]: I0313 10:26:03.323705 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zzmpk" podStartSLOduration=2.792312188 podStartE2EDuration="10.323683454s" podCreationTimestamp="2026-03-13 10:25:53 +0000 UTC" firstStartedPulling="2026-03-13 10:25:55.199978635 +0000 UTC m=+4395.949893312" lastFinishedPulling="2026-03-13 10:26:02.731349891 +0000 UTC m=+4403.481264578" observedRunningTime="2026-03-13 10:26:03.31557156 +0000 UTC m=+4404.065486247" watchObservedRunningTime="2026-03-13 10:26:03.323683454 +0000 UTC m=+4404.073598141" Mar 13 10:26:03 crc kubenswrapper[4930]: I0313 10:26:03.358797 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556626-bfbhx" podStartSLOduration=2.082767836 podStartE2EDuration="3.358776467s" podCreationTimestamp="2026-03-13 10:26:00 +0000 UTC" firstStartedPulling="2026-03-13 10:26:01.065729328 +0000 UTC m=+4401.815644005" lastFinishedPulling="2026-03-13 10:26:02.341737969 +0000 UTC m=+4403.091652636" observedRunningTime="2026-03-13 10:26:03.34697615 +0000 UTC m=+4404.096890837" watchObservedRunningTime="2026-03-13 10:26:03.358776467 +0000 UTC m=+4404.108691144" Mar 13 10:26:03 crc kubenswrapper[4930]: I0313 10:26:03.775630 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:26:03 crc kubenswrapper[4930]: I0313 10:26:03.775881 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:26:04 crc kubenswrapper[4930]: I0313 10:26:04.306615 4930 generic.go:334] "Generic (PLEG): container finished" podID="44ba6b14-c48d-443a-bb78-cf3e560f3a46" containerID="c62ea15f39c602bfe3561c09b828e926f13747cd5106b1bae48468bf120409a3" exitCode=0 Mar 13 10:26:04 crc kubenswrapper[4930]: I0313 10:26:04.306939 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556626-bfbhx" event={"ID":"44ba6b14-c48d-443a-bb78-cf3e560f3a46","Type":"ContainerDied","Data":"c62ea15f39c602bfe3561c09b828e926f13747cd5106b1bae48468bf120409a3"} Mar 13 10:26:05 crc kubenswrapper[4930]: I0313 10:26:05.035602 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zzmpk" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerName="registry-server" probeResult="failure" output=< Mar 13 10:26:05 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:26:05 crc kubenswrapper[4930]: > Mar 13 10:26:05 crc kubenswrapper[4930]: I0313 10:26:05.971835 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:26:05 crc kubenswrapper[4930]: E0313 10:26:05.972518 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:26:06 crc kubenswrapper[4930]: I0313 10:26:06.330929 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556626-bfbhx" event={"ID":"44ba6b14-c48d-443a-bb78-cf3e560f3a46","Type":"ContainerDied","Data":"59915d0e4c274c8d763699a4feb3cdaae8868caa77291c92f4815b3b832242ed"} Mar 13 10:26:06 crc kubenswrapper[4930]: I0313 10:26:06.330980 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59915d0e4c274c8d763699a4feb3cdaae8868caa77291c92f4815b3b832242ed" Mar 13 10:26:06 crc kubenswrapper[4930]: I0313 10:26:06.508136 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556626-bfbhx" Mar 13 10:26:06 crc kubenswrapper[4930]: I0313 10:26:06.624795 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55w4n\" (UniqueName: \"kubernetes.io/projected/44ba6b14-c48d-443a-bb78-cf3e560f3a46-kube-api-access-55w4n\") pod \"44ba6b14-c48d-443a-bb78-cf3e560f3a46\" (UID: \"44ba6b14-c48d-443a-bb78-cf3e560f3a46\") " Mar 13 10:26:06 crc kubenswrapper[4930]: I0313 10:26:06.637697 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44ba6b14-c48d-443a-bb78-cf3e560f3a46-kube-api-access-55w4n" (OuterVolumeSpecName: "kube-api-access-55w4n") pod "44ba6b14-c48d-443a-bb78-cf3e560f3a46" (UID: "44ba6b14-c48d-443a-bb78-cf3e560f3a46"). InnerVolumeSpecName "kube-api-access-55w4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:06 crc kubenswrapper[4930]: I0313 10:26:06.729318 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55w4n\" (UniqueName: \"kubernetes.io/projected/44ba6b14-c48d-443a-bb78-cf3e560f3a46-kube-api-access-55w4n\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:07 crc kubenswrapper[4930]: I0313 10:26:07.342718 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556626-bfbhx" Mar 13 10:26:07 crc kubenswrapper[4930]: I0313 10:26:07.573904 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556620-7zrwd"] Mar 13 10:26:07 crc kubenswrapper[4930]: I0313 10:26:07.588906 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556620-7zrwd"] Mar 13 10:26:07 crc kubenswrapper[4930]: I0313 10:26:07.984841 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0715803e-233a-45bd-9549-a619e4767274" path="/var/lib/kubelet/pods/0715803e-233a-45bd-9549-a619e4767274/volumes" Mar 13 10:26:14 crc kubenswrapper[4930]: I0313 10:26:14.821933 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zzmpk" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerName="registry-server" probeResult="failure" output=< Mar 13 10:26:14 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:26:14 crc kubenswrapper[4930]: > Mar 13 10:26:18 crc kubenswrapper[4930]: I0313 10:26:18.971347 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:26:18 crc kubenswrapper[4930]: E0313 10:26:18.972547 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:26:24 crc kubenswrapper[4930]: I0313 10:26:24.826425 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zzmpk" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerName="registry-server" probeResult="failure" output=< Mar 13 10:26:24 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:26:24 crc kubenswrapper[4930]: > Mar 13 10:26:29 crc kubenswrapper[4930]: I0313 10:26:29.970876 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:26:29 crc kubenswrapper[4930]: E0313 10:26:29.972003 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.065763 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Mar 13 10:26:33 crc kubenswrapper[4930]: E0313 10:26:33.067065 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44ba6b14-c48d-443a-bb78-cf3e560f3a46" containerName="oc" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.067081 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="44ba6b14-c48d-443a-bb78-cf3e560f3a46" containerName="oc" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.067353 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="44ba6b14-c48d-443a-bb78-cf3e560f3a46" containerName="oc" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.068645 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.072605 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.074095 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.074220 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.090698 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.094175 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4srqg" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.177574 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3b093601-3b3c-4541-ba63-bccaeedc6606-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.177909 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.177935 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b093601-3b3c-4541-ba63-bccaeedc6606-config-data\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.178010 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.178027 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b093601-3b3c-4541-ba63-bccaeedc6606-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.178071 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.178123 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b093601-3b3c-4541-ba63-bccaeedc6606-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.178181 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.178219 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpcf7\" (UniqueName: \"kubernetes.io/projected/3b093601-3b3c-4541-ba63-bccaeedc6606-kube-api-access-wpcf7\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.293306 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.293460 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpcf7\" (UniqueName: \"kubernetes.io/projected/3b093601-3b3c-4541-ba63-bccaeedc6606-kube-api-access-wpcf7\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.293595 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3b093601-3b3c-4541-ba63-bccaeedc6606-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.293661 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.293690 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b093601-3b3c-4541-ba63-bccaeedc6606-config-data\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.293786 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.293816 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b093601-3b3c-4541-ba63-bccaeedc6606-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.293872 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.293951 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b093601-3b3c-4541-ba63-bccaeedc6606-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.294643 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b093601-3b3c-4541-ba63-bccaeedc6606-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.295572 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b093601-3b3c-4541-ba63-bccaeedc6606-config-data\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.299736 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b093601-3b3c-4541-ba63-bccaeedc6606-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.302098 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.302234 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.302300 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3b093601-3b3c-4541-ba63-bccaeedc6606-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.319164 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.322978 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpcf7\" (UniqueName: \"kubernetes.io/projected/3b093601-3b3c-4541-ba63-bccaeedc6606-kube-api-access-wpcf7\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.330608 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.381738 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"tempest-tests-tempest\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.432757 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 13 10:26:33 crc kubenswrapper[4930]: I0313 10:26:33.937546 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 13 10:26:34 crc kubenswrapper[4930]: I0313 10:26:34.084924 4930 scope.go:117] "RemoveContainer" containerID="cdad3e9825ddafe8ba60cd1b7bfe35984314001bafddabd324fa132df2e34540" Mar 13 10:26:34 crc kubenswrapper[4930]: I0313 10:26:34.696321 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3b093601-3b3c-4541-ba63-bccaeedc6606","Type":"ContainerStarted","Data":"ffa0e3165df97abf8c31797782fe346109d8bfe39a7ece869adafae439631840"} Mar 13 10:26:34 crc kubenswrapper[4930]: I0313 10:26:34.824852 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zzmpk" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerName="registry-server" probeResult="failure" output=< Mar 13 10:26:34 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:26:34 crc kubenswrapper[4930]: > Mar 13 10:26:42 crc kubenswrapper[4930]: I0313 10:26:42.972050 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:26:42 crc kubenswrapper[4930]: E0313 10:26:42.973111 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:26:43 crc kubenswrapper[4930]: I0313 10:26:43.849960 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:26:43 crc kubenswrapper[4930]: I0313 10:26:43.909261 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:26:44 crc kubenswrapper[4930]: I0313 10:26:44.091910 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zzmpk"] Mar 13 10:26:45 crc kubenswrapper[4930]: I0313 10:26:45.824284 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zzmpk" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerName="registry-server" containerID="cri-o://da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01" gracePeriod=2 Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.667639 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.821908 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd737da-1d04-430b-8ee4-573fd5da59a3-catalog-content\") pod \"3cd737da-1d04-430b-8ee4-573fd5da59a3\" (UID: \"3cd737da-1d04-430b-8ee4-573fd5da59a3\") " Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.822086 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd737da-1d04-430b-8ee4-573fd5da59a3-utilities\") pod \"3cd737da-1d04-430b-8ee4-573fd5da59a3\" (UID: \"3cd737da-1d04-430b-8ee4-573fd5da59a3\") " Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.822131 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kxbm\" (UniqueName: \"kubernetes.io/projected/3cd737da-1d04-430b-8ee4-573fd5da59a3-kube-api-access-6kxbm\") pod \"3cd737da-1d04-430b-8ee4-573fd5da59a3\" (UID: \"3cd737da-1d04-430b-8ee4-573fd5da59a3\") " Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.823002 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cd737da-1d04-430b-8ee4-573fd5da59a3-utilities" (OuterVolumeSpecName: "utilities") pod "3cd737da-1d04-430b-8ee4-573fd5da59a3" (UID: "3cd737da-1d04-430b-8ee4-573fd5da59a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.835427 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cd737da-1d04-430b-8ee4-573fd5da59a3-kube-api-access-6kxbm" (OuterVolumeSpecName: "kube-api-access-6kxbm") pod "3cd737da-1d04-430b-8ee4-573fd5da59a3" (UID: "3cd737da-1d04-430b-8ee4-573fd5da59a3"). InnerVolumeSpecName "kube-api-access-6kxbm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.840488 4930 generic.go:334] "Generic (PLEG): container finished" podID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerID="da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01" exitCode=0 Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.840681 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzmpk" event={"ID":"3cd737da-1d04-430b-8ee4-573fd5da59a3","Type":"ContainerDied","Data":"da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01"} Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.840793 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzmpk" event={"ID":"3cd737da-1d04-430b-8ee4-573fd5da59a3","Type":"ContainerDied","Data":"bb5b5e94ec31e0e4e441c91575d6001bc31cb457dc2988406b5e8a717363a869"} Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.840960 4930 scope.go:117] "RemoveContainer" containerID="da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01" Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.841241 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzmpk" Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.921419 4930 scope.go:117] "RemoveContainer" containerID="b8cd4ecf098534fe7483d6c7617bb3a4258608ca7b6d528cf09606a735331458" Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.923840 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3cd737da-1d04-430b-8ee4-573fd5da59a3-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.923941 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kxbm\" (UniqueName: \"kubernetes.io/projected/3cd737da-1d04-430b-8ee4-573fd5da59a3-kube-api-access-6kxbm\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.951672 4930 scope.go:117] "RemoveContainer" containerID="167c36ac1141c75b6dd3f36900d6310aadf25272286e32c4e8aeacf1396169d6" Mar 13 10:26:46 crc kubenswrapper[4930]: I0313 10:26:46.962488 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3cd737da-1d04-430b-8ee4-573fd5da59a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3cd737da-1d04-430b-8ee4-573fd5da59a3" (UID: "3cd737da-1d04-430b-8ee4-573fd5da59a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:26:47 crc kubenswrapper[4930]: I0313 10:26:47.012860 4930 scope.go:117] "RemoveContainer" containerID="da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01" Mar 13 10:26:47 crc kubenswrapper[4930]: E0313 10:26:47.013287 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01\": container with ID starting with da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01 not found: ID does not exist" containerID="da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01" Mar 13 10:26:47 crc kubenswrapper[4930]: I0313 10:26:47.013322 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01"} err="failed to get container status \"da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01\": rpc error: code = NotFound desc = could not find container \"da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01\": container with ID starting with da591d85c56f9c36190fb5023b656a2fc6f09bd901ca67cd5dd8e8360000ee01 not found: ID does not exist" Mar 13 10:26:47 crc kubenswrapper[4930]: I0313 10:26:47.013343 4930 scope.go:117] "RemoveContainer" containerID="b8cd4ecf098534fe7483d6c7617bb3a4258608ca7b6d528cf09606a735331458" Mar 13 10:26:47 crc kubenswrapper[4930]: E0313 10:26:47.013920 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8cd4ecf098534fe7483d6c7617bb3a4258608ca7b6d528cf09606a735331458\": container with ID starting with b8cd4ecf098534fe7483d6c7617bb3a4258608ca7b6d528cf09606a735331458 not found: ID does not exist" containerID="b8cd4ecf098534fe7483d6c7617bb3a4258608ca7b6d528cf09606a735331458" Mar 13 10:26:47 crc kubenswrapper[4930]: I0313 10:26:47.013951 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8cd4ecf098534fe7483d6c7617bb3a4258608ca7b6d528cf09606a735331458"} err="failed to get container status \"b8cd4ecf098534fe7483d6c7617bb3a4258608ca7b6d528cf09606a735331458\": rpc error: code = NotFound desc = could not find container \"b8cd4ecf098534fe7483d6c7617bb3a4258608ca7b6d528cf09606a735331458\": container with ID starting with b8cd4ecf098534fe7483d6c7617bb3a4258608ca7b6d528cf09606a735331458 not found: ID does not exist" Mar 13 10:26:47 crc kubenswrapper[4930]: I0313 10:26:47.013969 4930 scope.go:117] "RemoveContainer" containerID="167c36ac1141c75b6dd3f36900d6310aadf25272286e32c4e8aeacf1396169d6" Mar 13 10:26:47 crc kubenswrapper[4930]: E0313 10:26:47.016109 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"167c36ac1141c75b6dd3f36900d6310aadf25272286e32c4e8aeacf1396169d6\": container with ID starting with 167c36ac1141c75b6dd3f36900d6310aadf25272286e32c4e8aeacf1396169d6 not found: ID does not exist" containerID="167c36ac1141c75b6dd3f36900d6310aadf25272286e32c4e8aeacf1396169d6" Mar 13 10:26:47 crc kubenswrapper[4930]: I0313 10:26:47.016239 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"167c36ac1141c75b6dd3f36900d6310aadf25272286e32c4e8aeacf1396169d6"} err="failed to get container status \"167c36ac1141c75b6dd3f36900d6310aadf25272286e32c4e8aeacf1396169d6\": rpc error: code = NotFound desc = could not find container \"167c36ac1141c75b6dd3f36900d6310aadf25272286e32c4e8aeacf1396169d6\": container with ID starting with 167c36ac1141c75b6dd3f36900d6310aadf25272286e32c4e8aeacf1396169d6 not found: ID does not exist" Mar 13 10:26:47 crc kubenswrapper[4930]: I0313 10:26:47.026796 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3cd737da-1d04-430b-8ee4-573fd5da59a3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:26:47 crc kubenswrapper[4930]: I0313 10:26:47.190782 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zzmpk"] Mar 13 10:26:47 crc kubenswrapper[4930]: I0313 10:26:47.203822 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zzmpk"] Mar 13 10:26:47 crc kubenswrapper[4930]: I0313 10:26:47.986827 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" path="/var/lib/kubelet/pods/3cd737da-1d04-430b-8ee4-573fd5da59a3/volumes" Mar 13 10:26:56 crc kubenswrapper[4930]: I0313 10:26:56.971043 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:26:56 crc kubenswrapper[4930]: E0313 10:26:56.971828 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:27:10 crc kubenswrapper[4930]: E0313 10:27:10.322848 4930 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Mar 13 10:27:10 crc kubenswrapper[4930]: E0313 10:27:10.328154 4930 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wpcf7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(3b093601-3b3c-4541-ba63-bccaeedc6606): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 13 10:27:10 crc kubenswrapper[4930]: E0313 10:27:10.329682 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="3b093601-3b3c-4541-ba63-bccaeedc6606" Mar 13 10:27:10 crc kubenswrapper[4930]: I0313 10:27:10.971463 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:27:10 crc kubenswrapper[4930]: E0313 10:27:10.971788 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:27:11 crc kubenswrapper[4930]: E0313 10:27:11.224381 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="3b093601-3b3c-4541-ba63-bccaeedc6606" Mar 13 10:27:25 crc kubenswrapper[4930]: I0313 10:27:25.531954 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 13 10:27:25 crc kubenswrapper[4930]: I0313 10:27:25.971267 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:27:25 crc kubenswrapper[4930]: E0313 10:27:25.971964 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:27:27 crc kubenswrapper[4930]: I0313 10:27:27.385617 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3b093601-3b3c-4541-ba63-bccaeedc6606","Type":"ContainerStarted","Data":"b4090d891229754dfe821cb8773a39f79da55d7c41f0d185fcee1e308b3f0554"} Mar 13 10:27:27 crc kubenswrapper[4930]: I0313 10:27:27.413050 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.832524985 podStartE2EDuration="56.413030791s" podCreationTimestamp="2026-03-13 10:26:31 +0000 UTC" firstStartedPulling="2026-03-13 10:26:33.948057051 +0000 UTC m=+4434.697971728" lastFinishedPulling="2026-03-13 10:27:25.528562867 +0000 UTC m=+4486.278477534" observedRunningTime="2026-03-13 10:27:27.400586908 +0000 UTC m=+4488.150501585" watchObservedRunningTime="2026-03-13 10:27:27.413030791 +0000 UTC m=+4488.162945458" Mar 13 10:27:40 crc kubenswrapper[4930]: I0313 10:27:40.971193 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:27:40 crc kubenswrapper[4930]: E0313 10:27:40.972017 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:27:54 crc kubenswrapper[4930]: I0313 10:27:54.971375 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:27:54 crc kubenswrapper[4930]: E0313 10:27:54.972152 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.151337 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556628-tjjtx"] Mar 13 10:28:00 crc kubenswrapper[4930]: E0313 10:28:00.153424 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerName="extract-content" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.153544 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerName="extract-content" Mar 13 10:28:00 crc kubenswrapper[4930]: E0313 10:28:00.153638 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerName="extract-utilities" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.153648 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerName="extract-utilities" Mar 13 10:28:00 crc kubenswrapper[4930]: E0313 10:28:00.153731 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerName="registry-server" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.153763 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerName="registry-server" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.154775 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="3cd737da-1d04-430b-8ee4-573fd5da59a3" containerName="registry-server" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.157994 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556628-tjjtx" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.161151 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.161349 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.161410 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.164830 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556628-tjjtx"] Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.221588 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f5zzm\" (UniqueName: \"kubernetes.io/projected/7acc7f58-2b75-428a-a2af-ced979a020c4-kube-api-access-f5zzm\") pod \"auto-csr-approver-29556628-tjjtx\" (UID: \"7acc7f58-2b75-428a-a2af-ced979a020c4\") " pod="openshift-infra/auto-csr-approver-29556628-tjjtx" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.323856 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f5zzm\" (UniqueName: \"kubernetes.io/projected/7acc7f58-2b75-428a-a2af-ced979a020c4-kube-api-access-f5zzm\") pod \"auto-csr-approver-29556628-tjjtx\" (UID: \"7acc7f58-2b75-428a-a2af-ced979a020c4\") " pod="openshift-infra/auto-csr-approver-29556628-tjjtx" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.342713 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f5zzm\" (UniqueName: \"kubernetes.io/projected/7acc7f58-2b75-428a-a2af-ced979a020c4-kube-api-access-f5zzm\") pod \"auto-csr-approver-29556628-tjjtx\" (UID: \"7acc7f58-2b75-428a-a2af-ced979a020c4\") " pod="openshift-infra/auto-csr-approver-29556628-tjjtx" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.483517 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556628-tjjtx" Mar 13 10:28:00 crc kubenswrapper[4930]: I0313 10:28:00.983345 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556628-tjjtx"] Mar 13 10:28:01 crc kubenswrapper[4930]: I0313 10:28:01.769395 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556628-tjjtx" event={"ID":"7acc7f58-2b75-428a-a2af-ced979a020c4","Type":"ContainerStarted","Data":"7422e40f174643c775f1d074bb33bd6d63b5589a5011be2d97e3ac30b61e02cd"} Mar 13 10:28:03 crc kubenswrapper[4930]: I0313 10:28:03.806642 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556628-tjjtx" event={"ID":"7acc7f58-2b75-428a-a2af-ced979a020c4","Type":"ContainerStarted","Data":"a6887e77eea5a880226dcc70437a06941c941f6c635f7833e650eb8a0d4580d6"} Mar 13 10:28:03 crc kubenswrapper[4930]: I0313 10:28:03.835045 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556628-tjjtx" podStartSLOduration=2.39450541 podStartE2EDuration="3.835022926s" podCreationTimestamp="2026-03-13 10:28:00 +0000 UTC" firstStartedPulling="2026-03-13 10:28:00.998607317 +0000 UTC m=+4521.748521994" lastFinishedPulling="2026-03-13 10:28:02.439124833 +0000 UTC m=+4523.189039510" observedRunningTime="2026-03-13 10:28:03.825958268 +0000 UTC m=+4524.575872965" watchObservedRunningTime="2026-03-13 10:28:03.835022926 +0000 UTC m=+4524.584937603" Mar 13 10:28:04 crc kubenswrapper[4930]: I0313 10:28:04.826571 4930 generic.go:334] "Generic (PLEG): container finished" podID="7acc7f58-2b75-428a-a2af-ced979a020c4" containerID="a6887e77eea5a880226dcc70437a06941c941f6c635f7833e650eb8a0d4580d6" exitCode=0 Mar 13 10:28:04 crc kubenswrapper[4930]: I0313 10:28:04.826724 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556628-tjjtx" event={"ID":"7acc7f58-2b75-428a-a2af-ced979a020c4","Type":"ContainerDied","Data":"a6887e77eea5a880226dcc70437a06941c941f6c635f7833e650eb8a0d4580d6"} Mar 13 10:28:06 crc kubenswrapper[4930]: I0313 10:28:06.415582 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556628-tjjtx" Mar 13 10:28:06 crc kubenswrapper[4930]: I0313 10:28:06.493804 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f5zzm\" (UniqueName: \"kubernetes.io/projected/7acc7f58-2b75-428a-a2af-ced979a020c4-kube-api-access-f5zzm\") pod \"7acc7f58-2b75-428a-a2af-ced979a020c4\" (UID: \"7acc7f58-2b75-428a-a2af-ced979a020c4\") " Mar 13 10:28:06 crc kubenswrapper[4930]: I0313 10:28:06.501295 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7acc7f58-2b75-428a-a2af-ced979a020c4-kube-api-access-f5zzm" (OuterVolumeSpecName: "kube-api-access-f5zzm") pod "7acc7f58-2b75-428a-a2af-ced979a020c4" (UID: "7acc7f58-2b75-428a-a2af-ced979a020c4"). InnerVolumeSpecName "kube-api-access-f5zzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:28:06 crc kubenswrapper[4930]: I0313 10:28:06.597981 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f5zzm\" (UniqueName: \"kubernetes.io/projected/7acc7f58-2b75-428a-a2af-ced979a020c4-kube-api-access-f5zzm\") on node \"crc\" DevicePath \"\"" Mar 13 10:28:06 crc kubenswrapper[4930]: I0313 10:28:06.852670 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556628-tjjtx" event={"ID":"7acc7f58-2b75-428a-a2af-ced979a020c4","Type":"ContainerDied","Data":"7422e40f174643c775f1d074bb33bd6d63b5589a5011be2d97e3ac30b61e02cd"} Mar 13 10:28:06 crc kubenswrapper[4930]: I0313 10:28:06.852709 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7422e40f174643c775f1d074bb33bd6d63b5589a5011be2d97e3ac30b61e02cd" Mar 13 10:28:06 crc kubenswrapper[4930]: I0313 10:28:06.852809 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556628-tjjtx" Mar 13 10:28:06 crc kubenswrapper[4930]: I0313 10:28:06.906311 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556622-fn42m"] Mar 13 10:28:06 crc kubenswrapper[4930]: I0313 10:28:06.917966 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556622-fn42m"] Mar 13 10:28:07 crc kubenswrapper[4930]: I0313 10:28:07.985718 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fca6144f-3b4c-4d42-af91-3bcb2a4fd06d" path="/var/lib/kubelet/pods/fca6144f-3b4c-4d42-af91-3bcb2a4fd06d/volumes" Mar 13 10:28:08 crc kubenswrapper[4930]: I0313 10:28:08.972950 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:28:08 crc kubenswrapper[4930]: E0313 10:28:08.975206 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:28:19 crc kubenswrapper[4930]: I0313 10:28:19.986261 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:28:19 crc kubenswrapper[4930]: E0313 10:28:19.987616 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:28:31 crc kubenswrapper[4930]: I0313 10:28:31.974941 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:28:31 crc kubenswrapper[4930]: E0313 10:28:31.975703 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:28:34 crc kubenswrapper[4930]: I0313 10:28:34.235561 4930 scope.go:117] "RemoveContainer" containerID="3df03912507e46221ac936ddf7139e3c087745163dcd99663ea52953b0a86ea2" Mar 13 10:28:45 crc kubenswrapper[4930]: I0313 10:28:45.975496 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:28:45 crc kubenswrapper[4930]: E0313 10:28:45.981169 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:28:56 crc kubenswrapper[4930]: I0313 10:28:56.973566 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:28:56 crc kubenswrapper[4930]: E0313 10:28:56.974911 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:29:10 crc kubenswrapper[4930]: I0313 10:29:10.980240 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:29:10 crc kubenswrapper[4930]: E0313 10:29:10.983409 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.368128 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mt2n8"] Mar 13 10:29:18 crc kubenswrapper[4930]: E0313 10:29:18.372165 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7acc7f58-2b75-428a-a2af-ced979a020c4" containerName="oc" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.372187 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="7acc7f58-2b75-428a-a2af-ced979a020c4" containerName="oc" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.374293 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="7acc7f58-2b75-428a-a2af-ced979a020c4" containerName="oc" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.380883 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.462322 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36c913ba-902f-480a-a1ba-ae8b95faa2d7-utilities\") pod \"community-operators-mt2n8\" (UID: \"36c913ba-902f-480a-a1ba-ae8b95faa2d7\") " pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.462401 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36c913ba-902f-480a-a1ba-ae8b95faa2d7-catalog-content\") pod \"community-operators-mt2n8\" (UID: \"36c913ba-902f-480a-a1ba-ae8b95faa2d7\") " pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.462494 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7v2z\" (UniqueName: \"kubernetes.io/projected/36c913ba-902f-480a-a1ba-ae8b95faa2d7-kube-api-access-q7v2z\") pod \"community-operators-mt2n8\" (UID: \"36c913ba-902f-480a-a1ba-ae8b95faa2d7\") " pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.524575 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mt2n8"] Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.564604 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36c913ba-902f-480a-a1ba-ae8b95faa2d7-utilities\") pod \"community-operators-mt2n8\" (UID: \"36c913ba-902f-480a-a1ba-ae8b95faa2d7\") " pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.564660 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36c913ba-902f-480a-a1ba-ae8b95faa2d7-catalog-content\") pod \"community-operators-mt2n8\" (UID: \"36c913ba-902f-480a-a1ba-ae8b95faa2d7\") " pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.564725 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7v2z\" (UniqueName: \"kubernetes.io/projected/36c913ba-902f-480a-a1ba-ae8b95faa2d7-kube-api-access-q7v2z\") pod \"community-operators-mt2n8\" (UID: \"36c913ba-902f-480a-a1ba-ae8b95faa2d7\") " pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.581485 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/36c913ba-902f-480a-a1ba-ae8b95faa2d7-utilities\") pod \"community-operators-mt2n8\" (UID: \"36c913ba-902f-480a-a1ba-ae8b95faa2d7\") " pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.582237 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/36c913ba-902f-480a-a1ba-ae8b95faa2d7-catalog-content\") pod \"community-operators-mt2n8\" (UID: \"36c913ba-902f-480a-a1ba-ae8b95faa2d7\") " pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.629630 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7v2z\" (UniqueName: \"kubernetes.io/projected/36c913ba-902f-480a-a1ba-ae8b95faa2d7-kube-api-access-q7v2z\") pod \"community-operators-mt2n8\" (UID: \"36c913ba-902f-480a-a1ba-ae8b95faa2d7\") " pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:18 crc kubenswrapper[4930]: I0313 10:29:18.769609 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:20 crc kubenswrapper[4930]: I0313 10:29:20.142097 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mt2n8"] Mar 13 10:29:20 crc kubenswrapper[4930]: I0313 10:29:20.711407 4930 generic.go:334] "Generic (PLEG): container finished" podID="36c913ba-902f-480a-a1ba-ae8b95faa2d7" containerID="a8d9aedeb3d234aca7bd49d5550aebf5b699784a38dc87abf87013b10e93f3e4" exitCode=0 Mar 13 10:29:20 crc kubenswrapper[4930]: I0313 10:29:20.711482 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mt2n8" event={"ID":"36c913ba-902f-480a-a1ba-ae8b95faa2d7","Type":"ContainerDied","Data":"a8d9aedeb3d234aca7bd49d5550aebf5b699784a38dc87abf87013b10e93f3e4"} Mar 13 10:29:20 crc kubenswrapper[4930]: I0313 10:29:20.711749 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mt2n8" event={"ID":"36c913ba-902f-480a-a1ba-ae8b95faa2d7","Type":"ContainerStarted","Data":"2df48489dda403b5357c5b377008d86bcdcf39936bfff8fe57a83ac35063896e"} Mar 13 10:29:22 crc kubenswrapper[4930]: I0313 10:29:22.975468 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:29:22 crc kubenswrapper[4930]: E0313 10:29:22.978779 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:29:29 crc kubenswrapper[4930]: I0313 10:29:29.819014 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mt2n8" event={"ID":"36c913ba-902f-480a-a1ba-ae8b95faa2d7","Type":"ContainerStarted","Data":"79dddd826e5a93570e8039df99d4755c5db320ee76156b00cb7daace60291252"} Mar 13 10:29:33 crc kubenswrapper[4930]: I0313 10:29:33.883960 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mt2n8" event={"ID":"36c913ba-902f-480a-a1ba-ae8b95faa2d7","Type":"ContainerDied","Data":"79dddd826e5a93570e8039df99d4755c5db320ee76156b00cb7daace60291252"} Mar 13 10:29:33 crc kubenswrapper[4930]: I0313 10:29:33.884515 4930 generic.go:334] "Generic (PLEG): container finished" podID="36c913ba-902f-480a-a1ba-ae8b95faa2d7" containerID="79dddd826e5a93570e8039df99d4755c5db320ee76156b00cb7daace60291252" exitCode=0 Mar 13 10:29:35 crc kubenswrapper[4930]: I0313 10:29:35.912938 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mt2n8" event={"ID":"36c913ba-902f-480a-a1ba-ae8b95faa2d7","Type":"ContainerStarted","Data":"952b077644fa686e6c6affadcd3a6d31cf2d3a8aaddf10d7415b20253e26a23c"} Mar 13 10:29:35 crc kubenswrapper[4930]: I0313 10:29:35.973932 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:29:35 crc kubenswrapper[4930]: E0313 10:29:35.975372 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-724mj_openshift-machine-config-operator(22188dce-43d2-4c7e-aa9b-7090a71eeb06)\"" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" Mar 13 10:29:36 crc kubenswrapper[4930]: I0313 10:29:36.036140 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mt2n8" podStartSLOduration=4.142143853 podStartE2EDuration="18.034517752s" podCreationTimestamp="2026-03-13 10:29:18 +0000 UTC" firstStartedPulling="2026-03-13 10:29:20.713122006 +0000 UTC m=+4601.463036683" lastFinishedPulling="2026-03-13 10:29:34.605495905 +0000 UTC m=+4615.355410582" observedRunningTime="2026-03-13 10:29:36.002820044 +0000 UTC m=+4616.752734721" watchObservedRunningTime="2026-03-13 10:29:36.034517752 +0000 UTC m=+4616.784432429" Mar 13 10:29:38 crc kubenswrapper[4930]: I0313 10:29:38.771069 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:38 crc kubenswrapper[4930]: I0313 10:29:38.771865 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:29:39 crc kubenswrapper[4930]: I0313 10:29:39.981564 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mt2n8" podUID="36c913ba-902f-480a-a1ba-ae8b95faa2d7" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:39 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:29:39 crc kubenswrapper[4930]: > Mar 13 10:29:47 crc kubenswrapper[4930]: I0313 10:29:47.976502 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:29:49 crc kubenswrapper[4930]: I0313 10:29:49.048535 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"2c05a807e57dc5147123981c1da8fc8655db8bfe897a8f8333698a2f312225b8"} Mar 13 10:29:50 crc kubenswrapper[4930]: I0313 10:29:50.115863 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mt2n8" podUID="36c913ba-902f-480a-a1ba-ae8b95faa2d7" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:50 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:29:50 crc kubenswrapper[4930]: > Mar 13 10:29:50 crc kubenswrapper[4930]: I0313 10:29:50.368617 4930 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-mbxlj container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.9:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:50 crc kubenswrapper[4930]: I0313 10:29:50.370504 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" podUID="58959680-d407-4688-8309-0438de9ce070" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.9:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:50 crc kubenswrapper[4930]: I0313 10:29:50.684587 4930 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:50 crc kubenswrapper[4930]: I0313 10:29:50.684654 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:50 crc kubenswrapper[4930]: I0313 10:29:50.837953 4930 patch_prober.go:28] interesting pod/console-7656bb8967-5vbb4 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.144:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:50 crc kubenswrapper[4930]: I0313 10:29:50.838013 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-7656bb8967-5vbb4" podUID="91a9c833-bbb1-4341-899b-d922b4066dc0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.144:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:51 crc kubenswrapper[4930]: I0313 10:29:51.314689 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-kf9cp" podUID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:51 crc kubenswrapper[4930]: timeout: health rpc did not complete within 1s Mar 13 10:29:51 crc kubenswrapper[4930]: > Mar 13 10:29:51 crc kubenswrapper[4930]: I0313 10:29:51.320580 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:51 crc kubenswrapper[4930]: timeout: health rpc did not complete within 1s Mar 13 10:29:51 crc kubenswrapper[4930]: > Mar 13 10:29:51 crc kubenswrapper[4930]: I0313 10:29:51.320653 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:51 crc kubenswrapper[4930]: timeout: health rpc did not complete within 1s Mar 13 10:29:51 crc kubenswrapper[4930]: > Mar 13 10:29:51 crc kubenswrapper[4930]: I0313 10:29:51.320777 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-kf9cp" podUID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:51 crc kubenswrapper[4930]: timeout: health rpc did not complete within 1s Mar 13 10:29:51 crc kubenswrapper[4930]: > Mar 13 10:29:51 crc kubenswrapper[4930]: I0313 10:29:51.384606 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" podUID="fd7bda25-ce82-4cc8-bc80-ecd97b9622cc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:51 crc kubenswrapper[4930]: I0313 10:29:51.473873 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" podUID="ee0102ec-b840-47fb-bae7-4b6e04c231fa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:51 crc kubenswrapper[4930]: I0313 10:29:51.947723 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" podUID="046b75aa-f3a0-467d-8624-6a82aa748b32" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:51 crc kubenswrapper[4930]: I0313 10:29:51.947817 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" podUID="6c666666-f9ed-4183-9a92-e0334239aa3e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:52 crc kubenswrapper[4930]: I0313 10:29:52.397637 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" podUID="43277746-d9dc-4452-bcaf-8c4668ebf473" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:52 crc kubenswrapper[4930]: I0313 10:29:52.420419 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:52 crc kubenswrapper[4930]: I0313 10:29:52.420483 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:52 crc kubenswrapper[4930]: I0313 10:29:52.474454 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:52 crc kubenswrapper[4930]: I0313 10:29:52.474522 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.58:8083/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:53 crc kubenswrapper[4930]: I0313 10:29:53.397957 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" podUID="cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.96:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:53 crc kubenswrapper[4930]: I0313 10:29:53.562651 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" podUID="f338c47b-e688-4de8-8bd7-02be68a70ad3" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:53 crc kubenswrapper[4930]: I0313 10:29:53.562670 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" podUID="f338c47b-e688-4de8-8bd7-02be68a70ad3" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:53 crc kubenswrapper[4930]: I0313 10:29:53.696115 4930 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-zfz2k container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:53 crc kubenswrapper[4930]: I0313 10:29:53.696170 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" podUID="8f964d06-4f78-4e56-ac62-fc19a2020489" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:53 crc kubenswrapper[4930]: I0313 10:29:53.696332 4930 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-zfz2k container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:53 crc kubenswrapper[4930]: I0313 10:29:53.696355 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" podUID="8f964d06-4f78-4e56-ac62-fc19a2020489" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.104642 4930 patch_prober.go:28] interesting pod/thanos-querier-7c466dcfcb-k7psd container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.105029 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" podUID="951e59b6-d39a-4ec2-beda-4b6b8b3e20c8" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.329620 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" podUID="984e8970-7f0d-4ce8-87c2-56c787857225" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.127:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.329640 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" podUID="984e8970-7f0d-4ce8-87c2-56c787857225" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.127:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.377282 4930 patch_prober.go:28] interesting pod/controller-manager-7984f94f96-clbzb container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.377400 4930 patch_prober.go:28] interesting pod/controller-manager-7984f94f96-clbzb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.377346 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" podUID="73d513ce-8d01-4a5e-ad2c-a3f86c80668a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.377469 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" podUID="73d513ce-8d01-4a5e-ad2c-a3f86c80668a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.603586 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-krmnx" podUID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.603679 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-krmnx" podUID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.686680 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-7bb4cc7c98-9bhdc" podUID="aa56d2a5-5b3a-487c-9248-21ef5689562f" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.99:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.686676 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-7bb4cc7c98-9bhdc" podUID="aa56d2a5-5b3a-487c-9248-21ef5689562f" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.99:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:54 crc kubenswrapper[4930]: I0313 10:29:54.686685 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-krmnx" podUID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.097158 4930 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.097261 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.139733 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" podUID="bbf3e078-1f66-4bcc-a742-0835afcb3e79" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.139778 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" podUID="bbf3e078-1f66-4bcc-a742-0835afcb3e79" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.452818 4930 patch_prober.go:28] interesting pod/route-controller-manager-55d94fcf7b-v9cp9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.452882 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" podUID="23510207-60fc-496b-b907-224360e7e887" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.452818 4930 patch_prober.go:28] interesting pod/route-controller-manager-55d94fcf7b-v9cp9 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.452994 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" podUID="23510207-60fc-496b-b907-224360e7e887" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.759927 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-zrwjj" podUID="ab0b3cdd-cfe6-4d41-8406-18104b17c0fb" containerName="nmstate-handler" probeResult="failure" output="command timed out" Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.759931 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.760102 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.760485 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.760629 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.787106 4930 patch_prober.go:28] interesting pod/nmstate-webhook-5f558f5558-z6585 container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.90:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:55 crc kubenswrapper[4930]: I0313 10:29:55.787165 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" podUID="9fc4b697-3810-4744-8c3f-2285ee8399e8" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.90:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:56 crc kubenswrapper[4930]: I0313 10:29:56.061674 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" podUID="77b779cf-5653-46d1-bbad-37b5efbbca03" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.45:6080/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:56 crc kubenswrapper[4930]: I0313 10:29:56.061827 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" podUID="77b779cf-5653-46d1-bbad-37b5efbbca03" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.45:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:56 crc kubenswrapper[4930]: I0313 10:29:56.106634 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-hc6zf" podUID="7f1370aa-666b-439e-9411-ec5e70360091" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:56 crc kubenswrapper[4930]: I0313 10:29:56.147640 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-hc6zf" podUID="7f1370aa-666b-439e-9411-ec5e70360091" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:56 crc kubenswrapper[4930]: I0313 10:29:56.552984 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-5rhcc" podUID="08a4c763-07ef-4163-84ea-7a9bb39de0f1" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:56 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:29:56 crc kubenswrapper[4930]: > Mar 13 10:29:56 crc kubenswrapper[4930]: I0313 10:29:56.552985 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-5rhcc" podUID="08a4c763-07ef-4163-84ea-7a9bb39de0f1" containerName="registry-server" probeResult="failure" output=< Mar 13 10:29:56 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:29:56 crc kubenswrapper[4930]: > Mar 13 10:29:56 crc kubenswrapper[4930]: I0313 10:29:56.760746 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="3459cebe-06a7-428e-93ef-419677fbcb58" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:29:56 crc kubenswrapper[4930]: I0313 10:29:56.760800 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="3459cebe-06a7-428e-93ef-419677fbcb58" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:29:56 crc kubenswrapper[4930]: I0313 10:29:56.881177 4930 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-9vpj8 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:56 crc kubenswrapper[4930]: I0313 10:29:56.881255 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" podUID="15d7d52b-dc81-4400-bbce-353f00ff5103" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:56 crc kubenswrapper[4930]: I0313 10:29:56.954676 4930 patch_prober.go:28] interesting pod/loki-operator-controller-manager-7f7d54c9b6-mhwwb container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.50:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:56 crc kubenswrapper[4930]: I0313 10:29:56.954755 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" podUID="c17ab33f-40db-4679-8e37-e3c4d013ba29" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.50:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.013517 4930 patch_prober.go:28] interesting pod/logging-loki-distributor-9c6b6d984-xfqch container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.013599 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" podUID="e70aee0c-0c7e-428a-9fc7-c099088bd6a0" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.162154 4930 patch_prober.go:28] interesting pod/logging-loki-querier-6dcbdf8bb8-kg8m5 container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.162501 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" podUID="e5b31a3a-a3d4-4e96-9d6a-6554af218386" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.278351 4930 patch_prober.go:28] interesting pod/logging-loki-query-frontend-ff66c4dc9-l2bss container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.278460 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" podUID="f63fa034-9ed3-4f32-9fae-c12ca9658f38" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.56:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.418925 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.419007 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.419398 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.419506 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.474558 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.474639 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.58:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.675618 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zptn9 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.675682 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" podUID="28acf0f4-52a0-40f4-9a05-abbdcc79fb33" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.675743 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.675704 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.675867 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zptn9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.675897 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.676110 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.58:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.676129 4930 patch_prober.go:28] interesting pod/metrics-server-647f59f4f4-vnqd8 container/metrics-server namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.85:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.676184 4930 patch_prober.go:28] interesting pod/metrics-server-647f59f4f4-vnqd8 container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.85:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.676217 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" podUID="e4d5bf77-4258-4108-ad1b-f51fa45a3bab" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.85:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.676240 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" podUID="d12dd258-eed7-41ff-b2e9-f124e189e9aa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.676284 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" podUID="d12dd258-eed7-41ff-b2e9-f124e189e9aa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.676259 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" podUID="e4d5bf77-4258-4108-ad1b-f51fa45a3bab" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.85:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.676396 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" podUID="28acf0f4-52a0-40f4-9a05-abbdcc79fb33" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.716638 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.716701 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.758619 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.758704 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.765050 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.773112 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840626 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840704 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840738 4930 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-j8gcf container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840739 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840760 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" podUID="1b960305-f6c1-4892-bbc2-70cce9c21546" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840698 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840805 4930 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-j8gcf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840805 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840818 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" podUID="1b960305-f6c1-4892-bbc2-70cce9c21546" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840844 4930 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8krq5 container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840848 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840866 4930 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8krq5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840877 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" podUID="d60fb09e-fa42-482c-9228-d89352573e75" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840878 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840850 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.840855 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" podUID="d60fb09e-fa42-482c-9228-d89352573e75" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.954053 4930 patch_prober.go:28] interesting pod/monitoring-plugin-7f784d5d74-hnt2f container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.86:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:57 crc kubenswrapper[4930]: I0313 10:29:57.954111 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" podUID="e52e16c5-4758-47c8-b6a3-0c2998455bd3" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.86:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.012994 4930 patch_prober.go:28] interesting pod/logging-loki-distributor-9c6b6d984-xfqch container/loki-distributor namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.54:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.013089 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" podUID="e70aee0c-0c7e-428a-9fc7-c099088bd6a0" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.54:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.120498 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.120563 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.120568 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.120701 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.162176 4930 patch_prober.go:28] interesting pod/logging-loki-querier-6dcbdf8bb8-kg8m5 container/loki-querier namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.55:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.162263 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" podUID="e5b31a3a-a3d4-4e96-9d6a-6554af218386" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.55:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.278196 4930 patch_prober.go:28] interesting pod/logging-loki-query-frontend-ff66c4dc9-l2bss container/loki-query-frontend namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.56:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.278399 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" podUID="f63fa034-9ed3-4f32-9fae-c12ca9658f38" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.56:3101/loki/api/v1/status/buildinfo\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.294735 4930 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.61:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.294814 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="89f3b1e5-d063-4df4-93e1-5815a3ddcaf6" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.61:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.342571 4930 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.62:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.342659 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="13c07815-f24e-4318-abb4-10ce48cfd113" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.62:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.419111 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/gateway namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.57:8081/live\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.419220 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/live\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.419256 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/opa namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.57:8083/live\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.419389 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/live\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.474406 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/gateway namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.58:8081/live\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.474496 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.58:8081/live\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.474572 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/opa namespace/openshift-logging: Liveness probe status=failure output="Get \"https://10.217.0.58:8083/live\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.474654 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.58:8083/live\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.480767 4930 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.480844 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a004ffaa-7bcb-425c-acbe-3dbd6154101d" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.717654 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:58 crc kubenswrapper[4930]: I0313 10:29:58.717693 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/healthy\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:59 crc kubenswrapper[4930]: I0313 10:29:59.103817 4930 patch_prober.go:28] interesting pod/thanos-querier-7c466dcfcb-k7psd container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:59 crc kubenswrapper[4930]: I0313 10:29:59.103943 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" podUID="951e59b6-d39a-4ec2-beda-4b6b8b3e20c8" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:59 crc kubenswrapper[4930]: I0313 10:29:59.628467 4930 trace.go:236] Trace[2094536532]: "Calculate volume metrics of wal for pod openshift-logging/logging-loki-ingester-0" (13-Mar-2026 10:29:52.713) (total time: 6911ms): Mar 13 10:29:59 crc kubenswrapper[4930]: Trace[2094536532]: [6.911046735s] [6.911046735s] END Mar 13 10:29:59 crc kubenswrapper[4930]: I0313 10:29:59.996991 4930 patch_prober.go:28] interesting pod/oauth-openshift-69c75b6f68-wdg64 container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:29:59 crc kubenswrapper[4930]: I0313 10:29:59.997066 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" podUID="37ebb02d-bc69-46d5-b9ac-3f676826de28" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:29:59 crc kubenswrapper[4930]: I0313 10:29:59.997108 4930 patch_prober.go:28] interesting pod/oauth-openshift-69c75b6f68-wdg64 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:29:59.997177 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" podUID="37ebb02d-bc69-46d5-b9ac-3f676826de28" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.025186 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.025299 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.025404 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.025594 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.240765 4930 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-mdgnp container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.240870 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.240781 4930 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-mdgnp container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.240933 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.410646 4930 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-mbxlj container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.9:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.410727 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" podUID="58959680-d407-4688-8309-0438de9ce070" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.9:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.410783 4930 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-mbxlj container/perses-operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.9:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.410797 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" podUID="58959680-d407-4688-8309-0438de9ce070" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.9:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.435069 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="79701ff1-8a11-4f24-9ab6-21f037545d5f" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.23:8081/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.435232 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="79701ff1-8a11-4f24-9ab6-21f037545d5f" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.23:8080/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.684808 4930 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.684886 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.836716 4930 patch_prober.go:28] interesting pod/console-7656bb8967-5vbb4 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.144:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:00 crc kubenswrapper[4930]: I0313 10:30:00.836791 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-7656bb8967-5vbb4" podUID="91a9c833-bbb1-4341-899b-d922b4066dc0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.144:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.027667 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" podUID="77b779cf-5653-46d1-bbad-37b5efbbca03" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.45:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.523019 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" podUID="fd7bda25-ce82-4cc8-bc80-ecd97b9622cc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.523178 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" podUID="404966af-7f68-4667-b535-81136694bdb9" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.523297 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" podUID="2e04d1ac-077a-4dc5-91a4-c9b22c352ac2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.604751 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" podUID="4051662d-ae7a-4f38-9aa5-5eecab9693f3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.686645 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" podUID="2e04d1ac-077a-4dc5-91a4-c9b22c352ac2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.686657 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" podUID="ee0102ec-b840-47fb-bae7-4b6e04c231fa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.763373 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.763370 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.769989 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" podUID="4a2edceb-a605-4edd-8905-e35177c281c7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.770061 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" podUID="fd7bda25-ce82-4cc8-bc80-ecd97b9622cc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.810009 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" podUID="4051662d-ae7a-4f38-9aa5-5eecab9693f3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.893901 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" podUID="d2ea88fb-18e6-4297-bb18-48d0e1d566be" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.976601 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" podUID="ee0102ec-b840-47fb-bae7-4b6e04c231fa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:01 crc kubenswrapper[4930]: I0313 10:30:01.976692 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" podUID="47a82845-cf9f-4fc8-a35f-4548a9c7127e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.141620 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" podUID="0d606b70-8e59-4303-963c-54c0be0a3800" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.223688 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" podUID="4a2edceb-a605-4edd-8905-e35177c281c7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.305697 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" podUID="585f7875-1138-4fcb-b89a-0a5f208f04b7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.387661 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" podUID="561d468b-4138-4196-8652-38b9eb4519ea" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.387710 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" podUID="12a9ad60-3dff-443d-9f09-b57c4a373aa2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.387862 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" podUID="d2ea88fb-18e6-4297-bb18-48d0e1d566be" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.420570 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.421024 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.475290 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.475648 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.58:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.552643 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" podUID="47a82845-cf9f-4fc8-a35f-4548a9c7127e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.552707 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" podUID="5ede266e-f821-4c72-a408-16b4b71a483e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.552665 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" podUID="046b75aa-f3a0-467d-8624-6a82aa748b32" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.552895 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" podUID="5ede266e-f821-4c72-a408-16b4b71a483e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.552935 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" podUID="0d606b70-8e59-4303-963c-54c0be0a3800" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.552952 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" podUID="a28dc996-b000-4680-b9f1-d770e91f0cba" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.593737 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" podUID="7601eb51-d514-4a96-848c-af0eae32f54f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.634693 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" podUID="a28dc996-b000-4680-b9f1-d770e91f0cba" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.634712 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" podUID="7601eb51-d514-4a96-848c-af0eae32f54f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.716769 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" podUID="585f7875-1138-4fcb-b89a-0a5f208f04b7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.716799 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podUID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.124:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.798696 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" podUID="561d468b-4138-4196-8652-38b9eb4519ea" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.798705 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" podUID="c92ab5a7-1ad7-4584-9954-0047b879bc47" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.125:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.880599 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" podUID="43277746-d9dc-4452-bcaf-8c4668ebf473" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.880814 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.880898 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" podUID="6c666666-f9ed-4183-9a92-e0334239aa3e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.880919 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.880923 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.880965 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.58:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.880902 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" podUID="6c666666-f9ed-4183-9a92-e0334239aa3e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.881018 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" podUID="046b75aa-f3a0-467d-8624-6a82aa748b32" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.881326 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podUID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.124:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.881379 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" podUID="c92ab5a7-1ad7-4584-9954-0047b879bc47" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.125:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:02 crc kubenswrapper[4930]: I0313 10:30:02.881389 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" podUID="43277746-d9dc-4452-bcaf-8c4668ebf473" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.022810 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.022881 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.022920 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.022947 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.397668 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" podUID="cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.96:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.565554 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" podUID="f338c47b-e688-4de8-8bd7-02be68a70ad3" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.565646 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" podUID="f338c47b-e688-4de8-8bd7-02be68a70ad3" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.668516 4930 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-zfz2k container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": context deadline exceeded" start-of-body= Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.668577 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" podUID="8f964d06-4f78-4e56-ac62-fc19a2020489" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": context deadline exceeded" Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.668662 4930 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-zfz2k container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.668681 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" podUID="8f964d06-4f78-4e56-ac62-fc19a2020489" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.717597 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/healthy\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.718166 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.764847 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Mar 13 10:30:03 crc kubenswrapper[4930]: I0313 10:30:03.766550 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mt2n8" podUID="36c913ba-902f-480a-a1ba-ae8b95faa2d7" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.103512 4930 patch_prober.go:28] interesting pod/thanos-querier-7c466dcfcb-k7psd container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.103593 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" podUID="951e59b6-d39a-4ec2-beda-4b6b8b3e20c8" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.286613 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" podUID="984e8970-7f0d-4ce8-87c2-56c787857225" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.127:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.376610 4930 patch_prober.go:28] interesting pod/controller-manager-7984f94f96-clbzb container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.376685 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" podUID="73d513ce-8d01-4a5e-ad2c-a3f86c80668a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.376610 4930 patch_prober.go:28] interesting pod/controller-manager-7984f94f96-clbzb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.377758 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" podUID="73d513ce-8d01-4a5e-ad2c-a3f86c80668a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.604583 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-krmnx" podUID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.604667 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-krmnx" podUID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.687618 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-krmnx" podUID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.687668 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-7bb4cc7c98-9bhdc" podUID="aa56d2a5-5b3a-487c-9248-21ef5689562f" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.99:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.687724 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-7bb4cc7c98-9bhdc" podUID="aa56d2a5-5b3a-487c-9248-21ef5689562f" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.99:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.764869 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-kf9cp" podUID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:04 crc kubenswrapper[4930]: I0313 10:30:04.765813 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-kf9cp" podUID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.130705 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" podUID="bbf3e078-1f66-4bcc-a742-0835afcb3e79" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.130766 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" podUID="bbf3e078-1f66-4bcc-a742-0835afcb3e79" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.130814 4930 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.131118 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.452757 4930 patch_prober.go:28] interesting pod/route-controller-manager-55d94fcf7b-v9cp9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.452790 4930 patch_prober.go:28] interesting pod/route-controller-manager-55d94fcf7b-v9cp9 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.452833 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" podUID="23510207-60fc-496b-b907-224360e7e887" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.452873 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" podUID="23510207-60fc-496b-b907-224360e7e887" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.761252 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.761357 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.764851 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.764922 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.786578 4930 patch_prober.go:28] interesting pod/nmstate-webhook-5f558f5558-z6585 container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.90:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.786623 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" podUID="9fc4b697-3810-4744-8c3f-2285ee8399e8" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.90:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:05 crc kubenswrapper[4930]: I0313 10:30:05.926651 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" podUID="4f1e0b58-9978-423f-84bc-f974dc6347b4" containerName="hostpath-provisioner" probeResult="failure" output="Get \"http://10.217.0.40:9898/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.020554 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.020646 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.020653 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.020753 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.027563 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.027713 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.030492 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="openshift-config-operator" containerStatusID={"Type":"cri-o","ID":"96e5faa3c3871a454979909e86e16277494094e8bfbd8123b70f92326e00fa79"} pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" containerMessage="Container openshift-config-operator failed liveness probe, will be restarted" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.031131 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" containerID="cri-o://96e5faa3c3871a454979909e86e16277494094e8bfbd8123b70f92326e00fa79" gracePeriod=30 Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.146648 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-hc6zf" podUID="7f1370aa-666b-439e-9411-ec5e70360091" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.146673 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-hc6zf" podUID="7f1370aa-666b-439e-9411-ec5e70360091" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.317904 4930 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-hjd4q container/oauth-apiserver namespace/openshift-oauth-apiserver: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.317982 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" podUID="1294227b-6b51-4269-93e2-542650c9dcb8" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.761420 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-zrwjj" podUID="ab0b3cdd-cfe6-4d41-8406-18104b17c0fb" containerName="nmstate-handler" probeResult="failure" output="command timed out" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.761558 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="3459cebe-06a7-428e-93ef-419677fbcb58" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.762126 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="3459cebe-06a7-428e-93ef-419677fbcb58" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.762698 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.762795 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.763458 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.881537 4930 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-9vpj8 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.881630 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" podUID="15d7d52b-dc81-4400-bbce-353f00ff5103" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.994773 4930 patch_prober.go:28] interesting pod/loki-operator-controller-manager-7f7d54c9b6-mhwwb container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.50:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.994858 4930 patch_prober.go:28] interesting pod/loki-operator-controller-manager-7f7d54c9b6-mhwwb container/manager namespace/openshift-operators-redhat: Liveness probe status=failure output="Get \"http://10.217.0.50:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.994881 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" podUID="c17ab33f-40db-4679-8e37-e3c4d013ba29" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.50:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.994902 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" podUID="c17ab33f-40db-4679-8e37-e3c4d013ba29" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.50:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:06 crc kubenswrapper[4930]: I0313 10:30:06.994797 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" podUID="4f1e0b58-9978-423f-84bc-f974dc6347b4" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.013159 4930 patch_prober.go:28] interesting pod/logging-loki-distributor-9c6b6d984-xfqch container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.013248 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" podUID="e70aee0c-0c7e-428a-9fc7-c099088bd6a0" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.028390 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.028529 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.162101 4930 patch_prober.go:28] interesting pod/logging-loki-querier-6dcbdf8bb8-kg8m5 container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.162191 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" podUID="e5b31a3a-a3d4-4e96-9d6a-6554af218386" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.279792 4930 patch_prober.go:28] interesting pod/logging-loki-query-frontend-ff66c4dc9-l2bss container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.279861 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" podUID="f63fa034-9ed3-4f32-9fae-c12ca9658f38" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.56:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.419632 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.419681 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.419704 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.419752 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.475606 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.475684 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.58:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.592618 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" podUID="28acf0f4-52a0-40f4-9a05-abbdcc79fb33" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.592631 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zptn9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.592943 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.592747 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" podUID="d12dd258-eed7-41ff-b2e9-f124e189e9aa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.592639 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.593013 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.58:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.592682 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zptn9 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.593089 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.592721 4930 patch_prober.go:28] interesting pod/metrics-server-647f59f4f4-vnqd8 container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.85:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.593137 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" podUID="e4d5bf77-4258-4108-ad1b-f51fa45a3bab" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.85:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.738995 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.739074 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.739005 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.739218 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.765411 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-g6hb8" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.765523 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-g6hb8" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.766040 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-5rhcc" podUID="08a4c763-07ef-4163-84ea-7a9bb39de0f1" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.766193 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-5rhcc" podUID="08a4c763-07ef-4163-84ea-7a9bb39de0f1" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.821742 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.821816 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.821874 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.821812 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.821935 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.821950 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.821887 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.821964 4930 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-j8gcf container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.821963 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.822019 4930 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8krq5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.822035 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" podUID="1b960305-f6c1-4892-bbc2-70cce9c21546" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.821985 4930 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-j8gcf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.822058 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" podUID="d60fb09e-fa42-482c-9228-d89352573e75" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.822002 4930 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8krq5 container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.822089 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" podUID="1b960305-f6c1-4892-bbc2-70cce9c21546" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.822094 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" podUID="d60fb09e-fa42-482c-9228-d89352573e75" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.954468 4930 patch_prober.go:28] interesting pod/monitoring-plugin-7f784d5d74-hnt2f container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.86:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:07 crc kubenswrapper[4930]: I0313 10:30:07.954539 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" podUID="e52e16c5-4758-47c8-b6a3-0c2998455bd3" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.86:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.007163 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" podUID="4f1e0b58-9978-423f-84bc-f974dc6347b4" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.121738 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.121811 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.121760 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.121963 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.214716 4930 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-z6996 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.214773 4930 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-z6996 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.214812 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" podUID="95528b17-57d5-4044-8c19-48dd83168399" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.214836 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" podUID="95528b17-57d5-4044-8c19-48dd83168399" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.295242 4930 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.61:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.295303 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="89f3b1e5-d063-4df4-93e1-5815a3ddcaf6" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.61:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.343581 4930 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.62:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.343664 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="13c07815-f24e-4318-abb4-10ce48cfd113" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.62:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.481266 4930 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.481327 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a004ffaa-7bcb-425c-acbe-3dbd6154101d" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.718072 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.718121 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/healthy\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.718487 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.763894 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovsdb-server" probeResult="failure" output="command timed out" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.765018 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-62lp6" podUID="195e681c-d03e-42fb-8c90-b97ca4c380b2" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.765191 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovs-vswitchd" probeResult="failure" output="command timed out" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.765204 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovsdb-server" probeResult="failure" output="command timed out" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.765253 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/certified-operators-62lp6" podUID="195e681c-d03e-42fb-8c90-b97ca4c380b2" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:08 crc kubenswrapper[4930]: I0313 10:30:08.766339 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovs-vswitchd" probeResult="failure" output="command timed out" Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.061630 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.061692 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.103784 4930 patch_prober.go:28] interesting pod/thanos-querier-7c466dcfcb-k7psd container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.103861 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" podUID="951e59b6-d39a-4ec2-beda-4b6b8b3e20c8" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.762079 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.763234 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-central-agent" probeResult="failure" output="command timed out" Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.763315 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ceilometer-0" Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.768747 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ceilometer-central-agent" containerStatusID={"Type":"cri-o","ID":"1cd5e3669981d79173656007193a66857ed29e6d11519fa9bc1873410a276c0f"} pod="openstack/ceilometer-0" containerMessage="Container ceilometer-central-agent failed liveness probe, will be restarted" Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.772469 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-central-agent" containerID="cri-o://1cd5e3669981d79173656007193a66857ed29e6d11519fa9bc1873410a276c0f" gracePeriod=30 Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.836688 4930 patch_prober.go:28] interesting pod/console-7656bb8967-5vbb4 container/console namespace/openshift-console: Liveness probe status=failure output="Get \"https://10.217.0.144:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.836758 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/console-7656bb8967-5vbb4" podUID="91a9c833-bbb1-4341-899b-d922b4066dc0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.144:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.836825 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.839079 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="console" containerStatusID={"Type":"cri-o","ID":"f8d94b25364c5b095aa56613345a2b6e246aa3b132a9875afaede44395624bec"} pod="openshift-console/console-7656bb8967-5vbb4" containerMessage="Container console failed liveness probe, will be restarted" Mar 13 10:30:09 crc kubenswrapper[4930]: I0313 10:30:09.905286 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" podUID="4f1e0b58-9978-423f-84bc-f974dc6347b4" containerName="hostpath-provisioner" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:09.996240 4930 patch_prober.go:28] interesting pod/oauth-openshift-69c75b6f68-wdg64 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:09.996310 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" podUID="37ebb02d-bc69-46d5-b9ac-3f676826de28" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:09.997011 4930 patch_prober.go:28] interesting pod/oauth-openshift-69c75b6f68-wdg64 container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:09.997055 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" podUID="37ebb02d-bc69-46d5-b9ac-3f676826de28" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.242723 4930 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-mdgnp container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.243230 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.242782 4930 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-mdgnp container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.243383 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.369717 4930 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-mbxlj container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.9:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.370086 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" podUID="58959680-d407-4688-8309-0438de9ce070" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.9:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.370208 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.428112 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="79701ff1-8a11-4f24-9ab6-21f037545d5f" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.23:8080/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.428139 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="79701ff1-8a11-4f24-9ab6-21f037545d5f" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.23:8081/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.684250 4930 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.684585 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.684696 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.762091 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.836986 4930 patch_prober.go:28] interesting pod/console-7656bb8967-5vbb4 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.144:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.837082 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-7656bb8967-5vbb4" podUID="91a9c833-bbb1-4341-899b-d922b4066dc0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.144:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.837197 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 10:30:10 crc kubenswrapper[4930]: I0313 10:30:10.940918 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-api-0" podUID="d82ac246-17a4-4f81-acce-70aaa7d8f350" containerName="cinder-api" probeResult="failure" output="Get \"https://10.217.0.228:8776/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.060692 4930 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.060768 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.102756 4930 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.102828 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.103255 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" podUID="77b779cf-5653-46d1-bbad-37b5efbbca03" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.45:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.119677 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.119761 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.479792 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" podUID="fd7bda25-ce82-4cc8-bc80-ecd97b9622cc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.479805 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" podUID="404966af-7f68-4667-b535-81136694bdb9" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.105:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.479917 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.520717 4930 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-mbxlj container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.9:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.520811 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" podUID="58959680-d407-4688-8309-0438de9ce070" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.9:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.524161 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" podUID="404966af-7f68-4667-b535-81136694bdb9" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.561694 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" podUID="4051662d-ae7a-4f38-9aa5-5eecab9693f3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.603794 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" podUID="2e04d1ac-077a-4dc5-91a4-c9b22c352ac2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.603855 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" podUID="ee0102ec-b840-47fb-bae7-4b6e04c231fa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.604059 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.644645 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" podUID="4a2edceb-a605-4edd-8905-e35177c281c7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.685631 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" podUID="12a9ad60-3dff-443d-9f09-b57c4a373aa2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.726583 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" podUID="d2ea88fb-18e6-4297-bb18-48d0e1d566be" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.760410 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-zrwjj" podUID="ab0b3cdd-cfe6-4d41-8406-18104b17c0fb" containerName="nmstate-handler" probeResult="failure" output="command timed out" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.767621 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" podUID="47a82845-cf9f-4fc8-a35f-4548a9c7127e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.808737 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" podUID="5ede266e-f821-4c72-a408-16b4b71a483e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.834483 4930 trace.go:236] Trace[883827431]: "Calculate volume metrics of storage for pod openshift-logging/logging-loki-ingester-0" (13-Mar-2026 10:29:59.631) (total time: 12198ms): Mar 13 10:30:11 crc kubenswrapper[4930]: Trace[883827431]: [12.198734746s] [12.198734746s] END Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.834489 4930 trace.go:236] Trace[962118761]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-1" (13-Mar-2026 10:30:01.455) (total time: 10375ms): Mar 13 10:30:11 crc kubenswrapper[4930]: Trace[962118761]: [10.375067853s] [10.375067853s] END Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.834505 4930 trace.go:236] Trace[1177834910]: "Calculate volume metrics of ovndbcluster-sb-etc-ovn for pod openstack/ovsdbserver-sb-0" (13-Mar-2026 10:29:57.825) (total time: 14004ms): Mar 13 10:30:11 crc kubenswrapper[4930]: Trace[1177834910]: [14.00490169s] [14.00490169s] END Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.849653 4930 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.849723 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.849761 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" podUID="0d606b70-8e59-4303-963c-54c0be0a3800" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.890777 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" podUID="a28dc996-b000-4680-b9f1-d770e91f0cba" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.890867 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.931757 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" podUID="585f7875-1138-4fcb-b89a-0a5f208f04b7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:11 crc kubenswrapper[4930]: I0313 10:30:11.973611 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" podUID="561d468b-4138-4196-8652-38b9eb4519ea" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.056638 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" podUID="046b75aa-f3a0-467d-8624-6a82aa748b32" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.056761 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.057011 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" podUID="6c666666-f9ed-4183-9a92-e0334239aa3e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.057076 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.275624 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" podUID="7601eb51-d514-4a96-848c-af0eae32f54f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.275636 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podUID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.124:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.321648 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" podUID="c92ab5a7-1ad7-4584-9954-0047b879bc47" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.125:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.395666 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" podUID="43277746-d9dc-4452-bcaf-8c4668ebf473" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.395994 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.419882 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.419940 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.419896 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.420890 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.474603 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.474674 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.58:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.474741 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.474762 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.58:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.521948 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" podUID="fd7bda25-ce82-4cc8-bc80-ecd97b9622cc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.645622 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" podUID="ee0102ec-b840-47fb-bae7-4b6e04c231fa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.920772 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:12 crc kubenswrapper[4930]: I0313 10:30:12.920833 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.139627 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" podUID="046b75aa-f3a0-467d-8624-6a82aa748b32" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.139949 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" podUID="6c666666-f9ed-4183-9a92-e0334239aa3e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.397701 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" podUID="cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.96:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.397822 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.438632 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" podUID="43277746-d9dc-4452-bcaf-8c4668ebf473" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.563648 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" podUID="f338c47b-e688-4de8-8bd7-02be68a70ad3" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.563738 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.563656 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" podUID="f338c47b-e688-4de8-8bd7-02be68a70ad3" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.565067 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.565881 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="webhook-server" containerStatusID={"Type":"cri-o","ID":"9130180554e9bb6c5586aab89960a91bc9711671a3b4b502faee937ff892eda1"} pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" containerMessage="Container webhook-server failed liveness probe, will be restarted" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.570604 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" podUID="f338c47b-e688-4de8-8bd7-02be68a70ad3" containerName="webhook-server" containerID="cri-o://9130180554e9bb6c5586aab89960a91bc9711671a3b4b502faee937ff892eda1" gracePeriod=2 Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.668615 4930 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-zfz2k container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.668673 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" podUID="8f964d06-4f78-4e56-ac62-fc19a2020489" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.668721 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.668723 4930 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-zfz2k container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.668807 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" podUID="8f964d06-4f78-4e56-ac62-fc19a2020489" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.669038 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.670224 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="prometheus-operator-admission-webhook" containerStatusID={"Type":"cri-o","ID":"4f86d8a5b91c950152ea5feebeab50790467d609ad31c9a05c0b392b87c01c02"} pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" containerMessage="Container prometheus-operator-admission-webhook failed liveness probe, will be restarted" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.670272 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" podUID="8f964d06-4f78-4e56-ac62-fc19a2020489" containerName="prometheus-operator-admission-webhook" containerID="cri-o://4f86d8a5b91c950152ea5feebeab50790467d609ad31c9a05c0b392b87c01c02" gracePeriod=30 Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.720728 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/healthy\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.763724 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.764260 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovsdb-server" probeResult="failure" output="command timed out" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.764380 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-controller-fghdj" podUID="2307735f-0b44-41ed-a622-74d62c05810c" containerName="ovn-controller" probeResult="failure" output="command timed out" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.765762 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovsdb-server" probeResult="failure" output="command timed out" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.765854 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mt2n8" podUID="36c913ba-902f-480a-a1ba-ae8b95faa2d7" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.765908 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovs-vswitchd" probeResult="failure" output="command timed out" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.765922 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovs-vswitchd" probeResult="failure" output="command timed out" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.765979 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-fghdj" podUID="2307735f-0b44-41ed-a622-74d62c05810c" containerName="ovn-controller" probeResult="failure" output="command timed out" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.926674 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" podUID="4f1e0b58-9978-423f-84bc-f974dc6347b4" containerName="hostpath-provisioner" probeResult="failure" output="Get \"http://10.217.0.40:9898/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.926761 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.928069 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="hostpath-provisioner" containerStatusID={"Type":"cri-o","ID":"8c960286a069fbf9bcd5f4c6337974f0f1c68a60d77a10e51960f73a826c9a7d"} pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" containerMessage="Container hostpath-provisioner failed liveness probe, will be restarted" Mar 13 10:30:13 crc kubenswrapper[4930]: I0313 10:30:13.928153 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" podUID="4f1e0b58-9978-423f-84bc-f974dc6347b4" containerName="hostpath-provisioner" containerID="cri-o://8c960286a069fbf9bcd5f4c6337974f0f1c68a60d77a10e51960f73a826c9a7d" gracePeriod=30 Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.020582 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.020711 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.104942 4930 patch_prober.go:28] interesting pod/thanos-querier-7c466dcfcb-k7psd container/kube-rbac-proxy-web namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.83:9091/-/healthy\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.105036 4930 patch_prober.go:28] interesting pod/thanos-querier-7c466dcfcb-k7psd container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.105118 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" podUID="951e59b6-d39a-4ec2-beda-4b6b8b3e20c8" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.105020 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" podUID="951e59b6-d39a-4ec2-beda-4b6b8b3e20c8" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.83:9091/-/healthy\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.328804 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" podUID="984e8970-7f0d-4ce8-87c2-56c787857225" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.127:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.328929 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.328819 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" podUID="984e8970-7f0d-4ce8-87c2-56c787857225" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.127:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.376165 4930 patch_prober.go:28] interesting pod/controller-manager-7984f94f96-clbzb container/controller-manager namespace/openshift-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.376226 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" podUID="73d513ce-8d01-4a5e-ad2c-a3f86c80668a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.376180 4930 patch_prober.go:28] interesting pod/controller-manager-7984f94f96-clbzb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.376280 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.376321 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" podUID="73d513ce-8d01-4a5e-ad2c-a3f86c80668a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.377812 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="controller-manager" containerStatusID={"Type":"cri-o","ID":"30bedad603a949a671f5e89c22fa86ae81a89a1bed0a3d726dda1b262d0cb66e"} pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" containerMessage="Container controller-manager failed liveness probe, will be restarted" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.377855 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" podUID="73d513ce-8d01-4a5e-ad2c-a3f86c80668a" containerName="controller-manager" containerID="cri-o://30bedad603a949a671f5e89c22fa86ae81a89a1bed0a3d726dda1b262d0cb66e" gracePeriod=30 Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.439695 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" podUID="cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.96:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.604717 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-krmnx" podUID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.604723 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-krmnx" podUID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.604826 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/frr-k8s-krmnx" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.604897 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-krmnx" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.613812 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="controller" containerStatusID={"Type":"cri-o","ID":"0b385e60f422e66ee531f65506a5ebd87f2b5f04b2d87a40d22a6bd2e3201027"} pod="metallb-system/frr-k8s-krmnx" containerMessage="Container controller failed liveness probe, will be restarted" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.614001 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-krmnx" podUID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerName="controller" containerID="cri-o://0b385e60f422e66ee531f65506a5ebd87f2b5f04b2d87a40d22a6bd2e3201027" gracePeriod=2 Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.688127 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-krmnx" podUID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.688172 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-7bb4cc7c98-9bhdc" podUID="aa56d2a5-5b3a-487c-9248-21ef5689562f" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.99:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.688222 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/frr-k8s-krmnx" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.688311 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.688297 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/controller-7bb4cc7c98-9bhdc" podUID="aa56d2a5-5b3a-487c-9248-21ef5689562f" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.99:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.688389 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.690180 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="controller" containerStatusID={"Type":"cri-o","ID":"54aac52da63da9f6e2134ae4af067569cf9545e3588d840ab16b2ee3eef8b684"} pod="metallb-system/controller-7bb4cc7c98-9bhdc" containerMessage="Container controller failed liveness probe, will be restarted" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.690254 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/controller-7bb4cc7c98-9bhdc" podUID="aa56d2a5-5b3a-487c-9248-21ef5689562f" containerName="controller" containerID="cri-o://54aac52da63da9f6e2134ae4af067569cf9545e3588d840ab16b2ee3eef8b684" gracePeriod=2 Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.764122 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-marketplace-kf9cp" podUID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.764212 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.766086 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-kf9cp" podUID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.766134 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85"} pod="openshift-marketplace/redhat-marketplace-kf9cp" containerMessage="Container registry-server failed liveness probe, will be restarted" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.766170 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.766177 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kf9cp" podUID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerName="registry-server" containerID="cri-o://6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85" gracePeriod=30 Mar 13 10:30:14 crc kubenswrapper[4930]: I0313 10:30:14.892090 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.132709 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" podUID="bbf3e078-1f66-4bcc-a742-0835afcb3e79" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.132736 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" podUID="bbf3e078-1f66-4bcc-a742-0835afcb3e79" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.132773 4930 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Liveness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.132794 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.132855 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.132968 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.132991 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.134512 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="frr-k8s-webhook-server" containerStatusID={"Type":"cri-o","ID":"9fe44cdc98892fa6b9a768d0a23c5444c75a778e1a00875feae8c0c303a524ed"} pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" containerMessage="Container frr-k8s-webhook-server failed liveness probe, will be restarted" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.134552 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" podUID="bbf3e078-1f66-4bcc-a742-0835afcb3e79" containerName="frr-k8s-webhook-server" containerID="cri-o://9fe44cdc98892fa6b9a768d0a23c5444c75a778e1a00875feae8c0c303a524ed" gracePeriod=10 Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.134920 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-scheduler" containerStatusID={"Type":"cri-o","ID":"02a86f4198baffedc25ef8a9b11068c3e64ca1cddf968d18f4f2c26520c0db44"} pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" containerMessage="Container kube-scheduler failed liveness probe, will be restarted" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.134991 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" containerID="cri-o://02a86f4198baffedc25ef8a9b11068c3e64ca1cddf968d18f4f2c26520c0db44" gracePeriod=30 Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.370683 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" podUID="984e8970-7f0d-4ce8-87c2-56c787857225" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.127:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.453472 4930 patch_prober.go:28] interesting pod/route-controller-manager-55d94fcf7b-v9cp9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.453538 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" podUID="23510207-60fc-496b-b907-224360e7e887" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.453556 4930 patch_prober.go:28] interesting pod/route-controller-manager-55d94fcf7b-v9cp9 container/route-controller-manager namespace/openshift-route-controller-manager: Liveness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.453613 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" podUID="23510207-60fc-496b-b907-224360e7e887" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.453674 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.455210 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="route-controller-manager" containerStatusID={"Type":"cri-o","ID":"b5434c352bda7c745416492632729de7378d748dea3b07785cf57f04f8ffb95d"} pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" containerMessage="Container route-controller-manager failed liveness probe, will be restarted" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.455251 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" podUID="23510207-60fc-496b-b907-224360e7e887" containerName="route-controller-manager" containerID="cri-o://b5434c352bda7c745416492632729de7378d748dea3b07785cf57f04f8ffb95d" gracePeriod=30 Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.691677 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-krmnx" podUID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerName="controller" probeResult="failure" output="Get \"http://127.0.0.1:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.691676 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.692288 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.733579 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/controller-7bb4cc7c98-9bhdc" podUID="aa56d2a5-5b3a-487c-9248-21ef5689562f" containerName="controller" probeResult="failure" output="Get \"http://10.217.0.99:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.761575 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.761575 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-galera-0" podUID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.761769 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.761802 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/openstack-galera-0" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.763301 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="galera" containerStatusID={"Type":"cri-o","ID":"c3c777a41680fafe3cd112f765e67426a68e7274d8368d2cff35e95608324d32"} pod="openstack/openstack-galera-0" containerMessage="Container galera failed liveness probe, will be restarted" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.766203 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.766281 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/redhat-operators-568x4" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.768043 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959"} pod="openshift-marketplace/redhat-operators-568x4" containerMessage="Container registry-server failed liveness probe, will be restarted" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.768087 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" containerID="cri-o://9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959" gracePeriod=30 Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.768577 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.768673 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-568x4" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.784936 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.785014 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.786544 4930 patch_prober.go:28] interesting pod/nmstate-webhook-5f558f5558-z6585 container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.90:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.786577 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" podUID="9fc4b697-3810-4744-8c3f-2285ee8399e8" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.90:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.786652 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.946639 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ca7da96e-e5ff-43fc-bca6-775f5aaba889" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.1.18:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.946612 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ca7da96e-e5ff-43fc-bca6-775f5aaba889" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.1.18:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.991742 4930 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:6443/readyz\": context deadline exceeded" start-of-body= Mar 13 10:30:15 crc kubenswrapper[4930]: I0313 10:30:15.991834 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/readyz\": context deadline exceeded" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.019256 4930 patch_prober.go:28] interesting pod/apiserver-76f77b778f-f2kpv container/openshift-apiserver namespace/openshift-apiserver: Liveness probe status=failure output="Get \"https://10.217.0.11:8443/livez?exclude=etcd\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.019329 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" podUID="45e7c581-937a-46b2-9501-b704239b106c" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.11:8443/livez?exclude=etcd\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.146778 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/speaker-hc6zf" podUID="7f1370aa-666b-439e-9411-ec5e70360091" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.146810 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-hc6zf" podUID="7f1370aa-666b-439e-9411-ec5e70360091" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.146863 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="metallb-system/speaker-hc6zf" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.146962 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hc6zf" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.149768 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="speaker" containerStatusID={"Type":"cri-o","ID":"b8fd9a8a2f4a99cfa5b81539c8e8c3d5bf5f01ce1422d2f280455d29d84073f6"} pod="metallb-system/speaker-hc6zf" containerMessage="Container speaker failed liveness probe, will be restarted" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.149828 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/speaker-hc6zf" podUID="7f1370aa-666b-439e-9411-ec5e70360091" containerName="speaker" containerID="cri-o://b8fd9a8a2f4a99cfa5b81539c8e8c3d5bf5f01ce1422d2f280455d29d84073f6" gracePeriod=2 Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.188694 4930 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.188729 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" podUID="bbf3e078-1f66-4bcc-a742-0835afcb3e79" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.188772 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.310993 4930 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-hjd4q container/oauth-apiserver namespace/openshift-oauth-apiserver: Liveness probe status=failure output="Get \"https://10.217.0.10:8443/livez?exclude=etcd\": context deadline exceeded" start-of-body= Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.311058 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" podUID="1294227b-6b51-4269-93e2-542650c9dcb8" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.10:8443/livez?exclude=etcd\": context deadline exceeded" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.319298 4930 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-hjd4q container/oauth-apiserver namespace/openshift-oauth-apiserver: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.319378 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-hjd4q" podUID="1294227b-6b51-4269-93e2-542650c9dcb8" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.10:8443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.742644 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7f55767858-jxftm" podUID="313c3b7e-2c50-4583-9dee-c141274c239a" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.210:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.742708 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-7f55767858-jxftm" podUID="313c3b7e-2c50-4583-9dee-c141274c239a" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.210:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.742774 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7f55767858-jxftm" podUID="313c3b7e-2c50-4583-9dee-c141274c239a" containerName="barbican-api" probeResult="failure" output="Get \"https://10.217.0.210:9311/healthcheck\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.742792 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/barbican-api-7f55767858-jxftm" podUID="313c3b7e-2c50-4583-9dee-c141274c239a" containerName="barbican-api-log" probeResult="failure" output="Get \"https://10.217.0.210:9311/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.760709 4930 patch_prober.go:28] interesting pod/image-registry-66df7c8f76-4sq9d container/registry namespace/openshift-image-registry: Readiness probe status=failure output="Get \"https://10.217.0.70:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.760757 4930 patch_prober.go:28] interesting pod/image-registry-66df7c8f76-4sq9d container/registry namespace/openshift-image-registry: Liveness probe status=failure output="Get \"https://10.217.0.70:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.760780 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" podUID="9027107f-0c82-4dd0-a8f1-01d09d646077" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.70:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.760804 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-image-registry/image-registry-66df7c8f76-4sq9d" podUID="9027107f-0c82-4dd0-a8f1-01d09d646077" containerName="registry" probeResult="failure" output="Get \"https://10.217.0.70:5000/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.762874 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.762874 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/openstack-cell1-galera-0" podUID="3459cebe-06a7-428e-93ef-419677fbcb58" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.762875 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-zrwjj" podUID="ab0b3cdd-cfe6-4d41-8406-18104b17c0fb" containerName="nmstate-handler" probeResult="failure" output="command timed out" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.762950 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.762963 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="3459cebe-06a7-428e-93ef-419677fbcb58" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.763113 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.763242 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.764496 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.764606 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="galera" containerStatusID={"Type":"cri-o","ID":"b85c79faf68f9744648c98621a97e4ae9f30c296fa782f54ee36a56172c622e1"} pod="openstack/openstack-cell1-galera-0" containerMessage="Container galera failed liveness probe, will be restarted" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.765085 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-engine-f9949f6d4-jldnd" podUID="c8e68bf4-ac94-446c-883a-758ad1dfdb89" containerName="heat-engine" probeResult="failure" output="command timed out" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.765926 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/heat-engine-f9949f6d4-jldnd" podUID="c8e68bf4-ac94-446c-883a-758ad1dfdb89" containerName="heat-engine" probeResult="failure" output="command timed out" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.787199 4930 patch_prober.go:28] interesting pod/nmstate-webhook-5f558f5558-z6585 container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.90:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.787268 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" podUID="9fc4b697-3810-4744-8c3f-2285ee8399e8" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.90:9443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.881169 4930 patch_prober.go:28] interesting pod/authentication-operator-69f744f599-9vpj8 container/authentication-operator namespace/openshift-authentication-operator: Liveness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.881244 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" podUID="15d7d52b-dc81-4400-bbce-353f00ff5103" containerName="authentication-operator" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.881299 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.882975 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="authentication-operator" containerStatusID={"Type":"cri-o","ID":"f8eac0b1a2476de135d617d21cdcefaae2507867e2a9d881bb3deb2366d37e5b"} pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" containerMessage="Container authentication-operator failed liveness probe, will be restarted" Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.883030 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" podUID="15d7d52b-dc81-4400-bbce-353f00ff5103" containerName="authentication-operator" containerID="cri-o://f8eac0b1a2476de135d617d21cdcefaae2507867e2a9d881bb3deb2366d37e5b" gracePeriod=30 Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.953733 4930 patch_prober.go:28] interesting pod/loki-operator-controller-manager-7f7d54c9b6-mhwwb container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.50:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:16 crc kubenswrapper[4930]: I0313 10:30:16.953820 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" podUID="c17ab33f-40db-4679-8e37-e3c4d013ba29" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.50:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:16.954422 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.021416 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.021522 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.025161 4930 patch_prober.go:28] interesting pod/logging-loki-distributor-9c6b6d984-xfqch container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.025206 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" podUID="e70aee0c-0c7e-428a-9fc7-c099088bd6a0" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.025286 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.189604 4930 patch_prober.go:28] interesting pod/logging-loki-querier-6dcbdf8bb8-kg8m5 container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.189651 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-hc6zf" podUID="7f1370aa-666b-439e-9411-ec5e70360091" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.189664 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" podUID="e5b31a3a-a3d4-4e96-9d6a-6554af218386" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.189859 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 10:30:17 crc kubenswrapper[4930]: E0313 10:30:17.190132 4930 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.278926 4930 patch_prober.go:28] interesting pod/logging-loki-query-frontend-ff66c4dc9-l2bss container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.279028 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" podUID="f63fa034-9ed3-4f32-9fae-c12ca9658f38" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.56:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.279147 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.419740 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.420035 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.419817 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.420186 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.430413 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" event={"ID":"f338c47b-e688-4de8-8bd7-02be68a70ad3","Type":"ContainerDied","Data":"9130180554e9bb6c5586aab89960a91bc9711671a3b4b502faee937ff892eda1"} Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.430739 4930 generic.go:334] "Generic (PLEG): container finished" podID="f338c47b-e688-4de8-8bd7-02be68a70ad3" containerID="9130180554e9bb6c5586aab89960a91bc9711671a3b4b502faee937ff892eda1" exitCode=137 Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.474080 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.474166 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.58:8081/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.675897 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zptn9 container/marketplace-operator namespace/openshift-marketplace: Liveness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.675941 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.675966 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.675966 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" podUID="28acf0f4-52a0-40f4-9a05-abbdcc79fb33" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676016 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676022 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.58:8083/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676037 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zptn9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676082 4930 patch_prober.go:28] interesting pod/metrics-server-647f59f4f4-vnqd8 container/metrics-server namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.85:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676094 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676100 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676118 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" podUID="e4d5bf77-4258-4108-ad1b-f51fa45a3bab" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.85:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676064 4930 patch_prober.go:28] interesting pod/metrics-server-647f59f4f4-vnqd8 container/metrics-server namespace/openshift-monitoring: Liveness probe status=failure output="Get \"https://10.217.0.85:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676154 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" podUID="e4d5bf77-4258-4108-ad1b-f51fa45a3bab" containerName="metrics-server" probeResult="failure" output="Get \"https://10.217.0.85:10250/livez\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676299 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" podUID="d12dd258-eed7-41ff-b2e9-f124e189e9aa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676341 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" podUID="d12dd258-eed7-41ff-b2e9-f124e189e9aa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676428 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" podUID="28acf0f4-52a0-40f4-9a05-abbdcc79fb33" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676539 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676577 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.676698 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.677687 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="marketplace-operator" containerStatusID={"Type":"cri-o","ID":"ef3feae5d76af1224b6c5c69d58de1290c3714a1add90314bbd94652c51a7da9"} pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" containerMessage="Container marketplace-operator failed liveness probe, will be restarted" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.677727 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" containerID="cri-o://ef3feae5d76af1224b6c5c69d58de1290c3714a1add90314bbd94652c51a7da9" gracePeriod=30 Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.678012 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="metrics-server" containerStatusID={"Type":"cri-o","ID":"5620f3a4a3a7c68ab03d8e44a6f4f2ed95ed87191aa5c9da5fa88541afd99535"} pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" containerMessage="Container metrics-server failed liveness probe, will be restarted" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.678036 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" podUID="e4d5bf77-4258-4108-ad1b-f51fa45a3bab" containerName="metrics-server" containerID="cri-o://5620f3a4a3a7c68ab03d8e44a6f4f2ed95ed87191aa5c9da5fa88541afd99535" gracePeriod=170 Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.758670 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.758748 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.758850 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.759259 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.759294 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.759328 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.760466 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="3459cebe-06a7-428e-93ef-419677fbcb58" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.760758 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="router" containerStatusID={"Type":"cri-o","ID":"0d1c91ff9aa1863e77bdd6ace633558b349856411c7ecae8822b1abe4009f2b6"} pod="openshift-ingress/router-default-5444994796-gc8fp" containerMessage="Container router failed liveness probe, will be restarted" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.760797 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" containerID="cri-o://0d1c91ff9aa1863e77bdd6ace633558b349856411c7ecae8822b1abe4009f2b6" gracePeriod=10 Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.762404 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.762509 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-5rhcc" podUID="08a4c763-07ef-4163-84ea-7a9bb39de0f1" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.762558 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.765450 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/certified-operators-62lp6" podUID="195e681c-d03e-42fb-8c90-b97ca4c380b2" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.765494 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/openstack-operator-index-5rhcc" podUID="08a4c763-07ef-4163-84ea-7a9bb39de0f1" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.765522 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/community-operators-g6hb8" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.765535 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.765902 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/community-operators-g6hb8" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.767470 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/certified-operators-62lp6" podUID="195e681c-d03e-42fb-8c90-b97ca4c380b2" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.841589 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.841642 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.841689 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.841940 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.841971 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.842018 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.842071 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.842097 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.842126 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.842802 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"401da0a18500414bd6d51283db6b432caf9b0add3d600d776c92ee41c49a1850"} pod="openshift-console/downloads-7954f5f757-22ldf" containerMessage="Container download-server failed liveness probe, will be restarted" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.842821 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="console-operator" containerStatusID={"Type":"cri-o","ID":"13aaec59d56c5086b74ff3d0f1745a7d4f6043bac526e05d5335c36cc41fe53f"} pod="openshift-console-operator/console-operator-58897d9998-w957v" containerMessage="Container console-operator failed liveness probe, will be restarted" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.842841 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" containerID="cri-o://401da0a18500414bd6d51283db6b432caf9b0add3d600d776c92ee41c49a1850" gracePeriod=2 Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.842845 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" containerID="cri-o://13aaec59d56c5086b74ff3d0f1745a7d4f6043bac526e05d5335c36cc41fe53f" gracePeriod=30 Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843189 4930 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-j8gcf container/olm-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843213 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" podUID="1b960305-f6c1-4892-bbc2-70cce9c21546" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843240 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843276 4930 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-j8gcf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843294 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" podUID="1b960305-f6c1-4892-bbc2-70cce9c21546" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843338 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843349 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843567 4930 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8krq5 container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843592 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" podUID="d60fb09e-fa42-482c-9228-d89352573e75" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843654 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843707 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843717 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843657 4930 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8krq5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843821 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" podUID="d60fb09e-fa42-482c-9228-d89352573e75" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.843903 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.844038 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="olm-operator" containerStatusID={"Type":"cri-o","ID":"7884531756cb3156917f38dc87d7796a7de383a57b50e29e1f711951f47dea05"} pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" containerMessage="Container olm-operator failed liveness probe, will be restarted" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.844073 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" podUID="1b960305-f6c1-4892-bbc2-70cce9c21546" containerName="olm-operator" containerID="cri-o://7884531756cb3156917f38dc87d7796a7de383a57b50e29e1f711951f47dea05" gracePeriod=30 Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.845275 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="packageserver" containerStatusID={"Type":"cri-o","ID":"d551980d1274b85a05cbf3ca612f7a73e8af7baef61f6c5b053db6dfdc5b6347"} pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" containerMessage="Container packageserver failed liveness probe, will be restarted" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.845530 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" podUID="d60fb09e-fa42-482c-9228-d89352573e75" containerName="packageserver" containerID="cri-o://d551980d1274b85a05cbf3ca612f7a73e8af7baef61f6c5b053db6dfdc5b6347" gracePeriod=30 Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.954707 4930 patch_prober.go:28] interesting pod/monitoring-plugin-7f784d5d74-hnt2f container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.86:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.954796 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" podUID="e52e16c5-4758-47c8-b6a3-0c2998455bd3" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.86:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.954877 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.996777 4930 patch_prober.go:28] interesting pod/loki-operator-controller-manager-7f7d54c9b6-mhwwb container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.50:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:17 crc kubenswrapper[4930]: I0313 10:30:17.996849 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" podUID="c17ab33f-40db-4679-8e37-e3c4d013ba29" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.50:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.026308 4930 patch_prober.go:28] interesting pod/logging-loki-distributor-9c6b6d984-xfqch container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.026379 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" podUID="e70aee0c-0c7e-428a-9fc7-c099088bd6a0" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.120977 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.121051 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": context deadline exceeded" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.121057 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.121119 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": context deadline exceeded" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.121422 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.121521 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.123069 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="catalog-operator" containerStatusID={"Type":"cri-o","ID":"e3e40cd070d4cdd05da72bed9064f18427537ff30f250f8f2e95be2d51084de2"} pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" containerMessage="Container catalog-operator failed liveness probe, will be restarted" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.123112 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" containerID="cri-o://e3e40cd070d4cdd05da72bed9064f18427537ff30f250f8f2e95be2d51084de2" gracePeriod=30 Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.214893 4930 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-z6996 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.214954 4930 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-z6996 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.215012 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" podUID="95528b17-57d5-4044-8c19-48dd83168399" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.215058 4930 patch_prober.go:28] interesting pod/logging-loki-querier-6dcbdf8bb8-kg8m5 container/loki-querier namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.215070 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" podUID="e5b31a3a-a3d4-4e96-9d6a-6554af218386" containerName="loki-querier" probeResult="failure" output="Get \"https://10.217.0.55:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.215333 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-z6996" podUID="95528b17-57d5-4044-8c19-48dd83168399" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.30:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.280078 4930 patch_prober.go:28] interesting pod/logging-loki-query-frontend-ff66c4dc9-l2bss container/loki-query-frontend namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.56:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.280135 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" podUID="f63fa034-9ed3-4f32-9fae-c12ca9658f38" containerName="loki-query-frontend" probeResult="failure" output="Get \"https://10.217.0.56:3101/loki/api/v1/status/buildinfo\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.294667 4930 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.61:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.294728 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="89f3b1e5-d063-4df4-93e1-5815a3ddcaf6" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.61:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.294833 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.342779 4930 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.62:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.342839 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="13c07815-f24e-4318-abb4-10ce48cfd113" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.62:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.342921 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.443386 4930 generic.go:334] "Generic (PLEG): container finished" podID="aa56d2a5-5b3a-487c-9248-21ef5689562f" containerID="54aac52da63da9f6e2134ae4af067569cf9545e3588d840ab16b2ee3eef8b684" exitCode=137 Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.443464 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-9bhdc" event={"ID":"aa56d2a5-5b3a-487c-9248-21ef5689562f","Type":"ContainerDied","Data":"54aac52da63da9f6e2134ae4af067569cf9545e3588d840ab16b2ee3eef8b684"} Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.444256 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="registry-server" containerStatusID={"Type":"cri-o","ID":"e1607208fa9fe34d7a2cd486255d55261b51f641995f51b7b3fd83a2ff83ad9c"} pod="openstack-operators/openstack-operator-index-5rhcc" containerMessage="Container registry-server failed liveness probe, will be restarted" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.444297 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-5rhcc" podUID="08a4c763-07ef-4163-84ea-7a9bb39de0f1" containerName="registry-server" containerID="cri-o://e1607208fa9fe34d7a2cd486255d55261b51f641995f51b7b3fd83a2ff83ad9c" gracePeriod=30 Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.480822 4930 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.480890 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a004ffaa-7bcb-425c-acbe-3dbd6154101d" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.480975 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.761366 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-zrwjj" podUID="ab0b3cdd-cfe6-4d41-8406-18104b17c0fb" containerName="nmstate-handler" probeResult="failure" output="command timed out" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.763865 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovsdb-server" probeResult="failure" output="command timed out" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.764081 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.764553 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovs-vswitchd" probeResult="failure" output="command timed out" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.764619 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.764774 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovs-vswitchd" probeResult="failure" output="command timed out" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.764974 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.765746 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ovs-vswitchd" containerStatusID={"Type":"cri-o","ID":"783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8"} pod="openstack/ovn-controller-ovs-zw9cp" containerMessage="Container ovs-vswitchd failed liveness probe, will be restarted" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.768943 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovsdb-server" probeResult="failure" output="command timed out" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.768994 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.800714 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.800824 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" podUID="d12dd258-eed7-41ff-b2e9-f124e189e9aa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.800882 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zptn9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.800905 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.800942 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/healthy\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.800973 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" podUID="28acf0f4-52a0-40f4-9a05-abbdcc79fb33" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.112:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.842643 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.842711 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.846746 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.846796 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.958150 4930 patch_prober.go:28] interesting pod/apiserver-76f77b778f-f2kpv container/openshift-apiserver namespace/openshift-apiserver: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz?exclude=etcd&exclude=etcd-readiness\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.958295 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-apiserver/apiserver-76f77b778f-f2kpv" podUID="45e7c581-937a-46b2-9501-b704239b106c" containerName="openshift-apiserver" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz?exclude=etcd&exclude=etcd-readiness\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.994595 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" podUID="4a79e474-ae50-4d4e-a809-29033d1357e9" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.1.35:8000/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:18 crc kubenswrapper[4930]: I0313 10:30:18.994620 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-575b8b7644-p4hwx" podUID="95d23f86-d860-42ac-9b8d-edf7854096d2" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.1.34:8004/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.076767 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/heat-cfnapi-6c996bbf9d-k5gp9" podUID="4a79e474-ae50-4d4e-a809-29033d1357e9" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.1.35:8000/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.076780 4930 patch_prober.go:28] interesting pod/monitoring-plugin-7f784d5d74-hnt2f container/monitoring-plugin namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.86:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.077208 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" podUID="e52e16c5-4758-47c8-b6a3-0c2998455bd3" containerName="monitoring-plugin" probeResult="failure" output="Get \"https://10.217.0.86:9443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.077081 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/heat-api-575b8b7644-p4hwx" podUID="95d23f86-d860-42ac-9b8d-edf7854096d2" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.1.34:8004/healthcheck\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.104216 4930 patch_prober.go:28] interesting pod/thanos-querier-7c466dcfcb-k7psd container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.104297 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" podUID="951e59b6-d39a-4ec2-beda-4b6b8b3e20c8" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.122319 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.122387 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.295912 4930 patch_prober.go:28] interesting pod/logging-loki-compactor-0 container/loki-compactor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.61:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.295983 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-compactor-0" podUID="89f3b1e5-d063-4df4-93e1-5815a3ddcaf6" containerName="loki-compactor" probeResult="failure" output="Get \"https://10.217.0.61:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.322099 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.322180 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.343904 4930 patch_prober.go:28] interesting pod/logging-loki-index-gateway-0 container/loki-index-gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.62:3101/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.344042 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-index-gateway-0" podUID="13c07815-f24e-4318-abb4-10ce48cfd113" containerName="loki-index-gateway" probeResult="failure" output="Get \"https://10.217.0.62:3101/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: E0313 10:30:19.378318 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:19 crc kubenswrapper[4930]: E0313 10:30:19.380344 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:19 crc kubenswrapper[4930]: E0313 10:30:19.381727 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:19 crc kubenswrapper[4930]: E0313 10:30:19.381769 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-kf9cp" podUID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerName="registry-server" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.462143 4930 generic.go:334] "Generic (PLEG): container finished" podID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerID="0b385e60f422e66ee531f65506a5ebd87f2b5f04b2d87a40d22a6bd2e3201027" exitCode=137 Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.462212 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerDied","Data":"0b385e60f422e66ee531f65506a5ebd87f2b5f04b2d87a40d22a6bd2e3201027"} Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.482329 4930 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.482412 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="a004ffaa-7bcb-425c-acbe-3dbd6154101d" containerName="loki-ingester" probeResult="failure" output="Get \"https://10.217.0.60:3101/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.685693 4930 patch_prober.go:28] interesting pod/openshift-kube-scheduler-crc container/kube-scheduler namespace/openshift-kube-scheduler: Readiness probe status=failure output="Get \"https://192.168.126.11:10259/healthz\": dial tcp 192.168.126.11:10259: connect: connection refused" start-of-body= Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.685765 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podUID="3dcd261975c3d6b9a6ad6367fd4facd3" containerName="kube-scheduler" probeResult="failure" output="Get \"https://192.168.126.11:10259/healthz\": dial tcp 192.168.126.11:10259: connect: connection refused" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.766457 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" podUID="c7123547-da61-46f0-9e1c-e54b075191bb" containerName="sbdb" probeResult="failure" output="command timed out" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.766564 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ovn-kubernetes/ovnkube-node-pgvdn" podUID="c7123547-da61-46f0-9e1c-e54b075191bb" containerName="nbdb" probeResult="failure" output="command timed out" Mar 13 10:30:19 crc kubenswrapper[4930]: E0313 10:30:19.976766 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:19 crc kubenswrapper[4930]: E0313 10:30:19.978820 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:19 crc kubenswrapper[4930]: E0313 10:30:19.980087 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:19 crc kubenswrapper[4930]: E0313 10:30:19.980128 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.982197 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="cert-manager/cert-manager-858654f9db-s4m2h" podUID="596bf37a-3597-46b7-a155-0e809f44b02a" containerName="cert-manager-controller" probeResult="failure" output="Get \"http://10.217.0.44:9403/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.996091 4930 patch_prober.go:28] interesting pod/oauth-openshift-69c75b6f68-wdg64 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.996121 4930 patch_prober.go:28] interesting pod/oauth-openshift-69c75b6f68-wdg64 container/oauth-openshift namespace/openshift-authentication: Liveness probe status=failure output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.996172 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" podUID="37ebb02d-bc69-46d5-b9ac-3f676826de28" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.996185 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" podUID="37ebb02d-bc69-46d5-b9ac-3f676826de28" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.996281 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.996314 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 10:30:19 crc kubenswrapper[4930]: I0313 10:30:19.998258 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="oauth-openshift" containerStatusID={"Type":"cri-o","ID":"34f191dfc2ed28b8f28030a7f3b536e07a821336bad5a53f8da29c19b66b5be9"} pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" containerMessage="Container oauth-openshift failed liveness probe, will be restarted" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.024609 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.024726 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.242638 4930 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-mdgnp container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.242702 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.242758 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.242650 4930 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-mdgnp container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.242836 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.242895 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.244414 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="operator" containerStatusID={"Type":"cri-o","ID":"168aa9a41e863b3e1cbb2cb4acc917874653648d080c859dd1cc0524ede1dedc"} pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" containerMessage="Container operator failed liveness probe, will be restarted" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.244497 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" containerName="operator" containerID="cri-o://168aa9a41e863b3e1cbb2cb4acc917874653648d080c859dd1cc0524ede1dedc" gracePeriod=30 Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.411661 4930 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-mbxlj container/perses-operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.9:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.411738 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" podUID="58959680-d407-4688-8309-0438de9ce070" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.9:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.411756 4930 patch_prober.go:28] interesting pod/perses-operator-5bf474d74f-mbxlj container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.9:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.411817 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" podUID="58959680-d407-4688-8309-0438de9ce070" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.9:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.427575 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="79701ff1-8a11-4f24-9ab6-21f037545d5f" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.23:8081/readyz\": context deadline exceeded" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.427590 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/kube-state-metrics-0" podUID="79701ff1-8a11-4f24-9ab6-21f037545d5f" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.23:8080/livez\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.428054 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/kube-state-metrics-0" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.428567 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.429745 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-state-metrics" containerStatusID={"Type":"cri-o","ID":"2bef53636a65aa4df31016d5e4a9dbca79a002bb165bc0e482564a6814cfcc43"} pod="openstack/kube-state-metrics-0" containerMessage="Container kube-state-metrics failed liveness probe, will be restarted" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.429804 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="79701ff1-8a11-4f24-9ab6-21f037545d5f" containerName="kube-state-metrics" containerID="cri-o://2bef53636a65aa4df31016d5e4a9dbca79a002bb165bc0e482564a6814cfcc43" gracePeriod=30 Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.480961 4930 generic.go:334] "Generic (PLEG): container finished" podID="7f1370aa-666b-439e-9411-ec5e70360091" containerID="b8fd9a8a2f4a99cfa5b81539c8e8c3d5bf5f01ce1422d2f280455d29d84073f6" exitCode=137 Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.481011 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hc6zf" event={"ID":"7f1370aa-666b-439e-9411-ec5e70360091","Type":"ContainerDied","Data":"b8fd9a8a2f4a99cfa5b81539c8e8c3d5bf5f01ce1422d2f280455d29d84073f6"} Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.836911 4930 patch_prober.go:28] interesting pod/console-7656bb8967-5vbb4 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.144:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.837000 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-7656bb8967-5vbb4" podUID="91a9c833-bbb1-4341-899b-d922b4066dc0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.144:8443/health\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.991245 4930 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:6443/livez?exclude=etcd\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:20 crc kubenswrapper[4930]: I0313 10:30:20.991323 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="Get \"https://192.168.126.11:6443/livez?exclude=etcd\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.020662 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-687f57d79b-nnnmp" podUID="77b779cf-5653-46d1-bbad-37b5efbbca03" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.45:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.020692 4930 patch_prober.go:28] interesting pod/oauth-openshift-69c75b6f68-wdg64 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.020787 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" podUID="37ebb02d-bc69-46d5-b9ac-3f676826de28" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.66:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.102667 4930 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.102764 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.143667 4930 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.143741 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.525772 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" podUID="fd7bda25-ce82-4cc8-bc80-ecd97b9622cc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.526021 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" podUID="404966af-7f68-4667-b535-81136694bdb9" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.526259 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.526057 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" podUID="2e04d1ac-077a-4dc5-91a4-c9b22c352ac2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.526744 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.607679 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" podUID="4051662d-ae7a-4f38-9aa5-5eecab9693f3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.690665 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" podUID="2e04d1ac-077a-4dc5-91a4-c9b22c352ac2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.690665 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" podUID="ee0102ec-b840-47fb-bae7-4b6e04c231fa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.717413 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_downloads-7954f5f757-22ldf_dc201d97-a254-45f7-8f60-f2c5dd726852/download-server/0.log" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.717493 4930 generic.go:334] "Generic (PLEG): container finished" podID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerID="401da0a18500414bd6d51283db6b432caf9b0add3d600d776c92ee41c49a1850" exitCode=137 Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.717529 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-22ldf" event={"ID":"dc201d97-a254-45f7-8f60-f2c5dd726852","Type":"ContainerDied","Data":"401da0a18500414bd6d51283db6b432caf9b0add3d600d776c92ee41c49a1850"} Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.760677 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-handler-zrwjj" podUID="ab0b3cdd-cfe6-4d41-8406-18104b17c0fb" containerName="nmstate-handler" probeResult="failure" output="command timed out" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.762837 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.763394 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.763476 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.765097 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="prometheus" containerStatusID={"Type":"cri-o","ID":"1687564f2e703b32a639e036942f5c7161b2659f59f0f165489b992552ee549a"} pod="openshift-monitoring/prometheus-k8s-0" containerMessage="Container prometheus failed liveness probe, will be restarted" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.765222 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" containerID="cri-o://1687564f2e703b32a639e036942f5c7161b2659f59f0f165489b992552ee549a" gracePeriod=600 Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.775904 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" podUID="4a2edceb-a605-4edd-8905-e35177c281c7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.776056 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" podUID="fd7bda25-ce82-4cc8-bc80-ecd97b9622cc" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.107:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.857915 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" podUID="12a9ad60-3dff-443d-9f09-b57c4a373aa2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.858366 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.858449 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" podUID="4051662d-ae7a-4f38-9aa5-5eecab9693f3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:21 crc kubenswrapper[4930]: I0313 10:30:21.858557 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.023654 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" podUID="ee0102ec-b840-47fb-bae7-4b6e04c231fa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.109:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.023662 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" podUID="5ede266e-f821-4c72-a408-16b4b71a483e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.230610 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" podUID="4a2edceb-a605-4edd-8905-e35177c281c7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.230720 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.316341 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" podUID="585f7875-1138-4fcb-b89a-0a5f208f04b7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.398861 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" podUID="12a9ad60-3dff-443d-9f09-b57c4a373aa2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.399054 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" podUID="561d468b-4138-4196-8652-38b9eb4519ea" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.445340 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" podUID="6c666666-f9ed-4183-9a92-e0334239aa3e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.527788 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" podUID="046b75aa-f3a0-467d-8624-6a82aa748b32" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.527801 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" podUID="d2ea88fb-18e6-4297-bb18-48d0e1d566be" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.113:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.527883 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.527929 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.527937 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.57:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.528461 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" podUID="47a82845-cf9f-4fc8-a35f-4548a9c7127e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.528503 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" podUID="47a82845-cf9f-4fc8-a35f-4548a9c7127e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.528524 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.528624 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" podUID="5ede266e-f821-4c72-a408-16b4b71a483e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.115:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.528708 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" podUID="0d606b70-8e59-4303-963c-54c0be0a3800" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.528736 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.528746 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" podUID="0d606b70-8e59-4303-963c-54c0be0a3800" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.528757 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/opa namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.528793 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.528812 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="opa" probeResult="failure" output="Get \"https://10.217.0.58:8083/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.528860 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" podUID="a28dc996-b000-4680-b9f1-d770e91f0cba" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.529042 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.653695 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podUID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.124:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.653694 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" podUID="a28dc996-b000-4680-b9f1-d770e91f0cba" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.653808 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" podUID="585f7875-1138-4fcb-b89a-0a5f208f04b7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.653939 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.755154 4930 generic.go:334] "Generic (PLEG): container finished" podID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerID="e3e40cd070d4cdd05da72bed9064f18427537ff30f250f8f2e95be2d51084de2" exitCode=0 Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.755434 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" event={"ID":"143f048a-f720-4d97-92ac-c9766a7e39ab","Type":"ContainerDied","Data":"e3e40cd070d4cdd05da72bed9064f18427537ff30f250f8f2e95be2d51084de2"} Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.762360 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-58897d9998-w957v_a2feeb9b-9255-4660-829f-23c296a2a1c3/console-operator/0.log" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.762422 4930 generic.go:334] "Generic (PLEG): container finished" podID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerID="13aaec59d56c5086b74ff3d0f1745a7d4f6043bac526e05d5335c36cc41fe53f" exitCode=1 Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.762542 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-w957v" event={"ID":"a2feeb9b-9255-4660-829f-23c296a2a1c3","Type":"ContainerDied","Data":"13aaec59d56c5086b74ff3d0f1745a7d4f6043bac526e05d5335c36cc41fe53f"} Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.765072 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-index-5rhcc" podUID="08a4c763-07ef-4163-84ea-7a9bb39de0f1" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:22 crc kubenswrapper[4930]: E0313 10:30:22.768229 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e1607208fa9fe34d7a2cd486255d55261b51f641995f51b7b3fd83a2ff83ad9c" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:22 crc kubenswrapper[4930]: E0313 10:30:22.772341 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e1607208fa9fe34d7a2cd486255d55261b51f641995f51b7b3fd83a2ff83ad9c" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.775690 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-9bhdc" event={"ID":"aa56d2a5-5b3a-487c-9248-21ef5689562f","Type":"ContainerStarted","Data":"33be98cd5508b2186ea1f8c9027c5ab6d1d35bdd364e62ec72c85bc10c25d95c"} Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.775890 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 10:30:22 crc kubenswrapper[4930]: E0313 10:30:22.777082 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e1607208fa9fe34d7a2cd486255d55261b51f641995f51b7b3fd83a2ff83ad9c" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:22 crc kubenswrapper[4930]: E0313 10:30:22.777129 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack-operators/openstack-operator-index-5rhcc" podUID="08a4c763-07ef-4163-84ea-7a9bb39de0f1" containerName="registry-server" Mar 13 10:30:22 crc kubenswrapper[4930]: E0313 10:30:22.778633 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e1607208fa9fe34d7a2cd486255d55261b51f641995f51b7b3fd83a2ff83ad9c" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:22 crc kubenswrapper[4930]: E0313 10:30:22.779786 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e1607208fa9fe34d7a2cd486255d55261b51f641995f51b7b3fd83a2ff83ad9c" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:22 crc kubenswrapper[4930]: E0313 10:30:22.780699 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="e1607208fa9fe34d7a2cd486255d55261b51f641995f51b7b3fd83a2ff83ad9c" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:22 crc kubenswrapper[4930]: E0313 10:30:22.780738 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack-operators/openstack-operator-index-5rhcc" podUID="08a4c763-07ef-4163-84ea-7a9bb39de0f1" containerName="registry-server" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.784593 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerStarted","Data":"f3e7ea93f2e849164e73dace4575eb4dd5314f047954d087eb62daabb2439513"} Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.784783 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-krmnx" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.785731 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="frr" containerStatusID={"Type":"cri-o","ID":"da0e90531521761cb303522a5a41ba4b7c379d88504e0716cbb1fa3ec99fdc76"} pod="metallb-system/frr-k8s-krmnx" containerMessage="Container frr failed liveness probe, will be restarted" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.785838 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="metallb-system/frr-k8s-krmnx" podUID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerName="frr" containerID="cri-o://da0e90531521761cb303522a5a41ba4b7c379d88504e0716cbb1fa3ec99fdc76" gracePeriod=2 Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.787759 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" event={"ID":"f338c47b-e688-4de8-8bd7-02be68a70ad3","Type":"ContainerStarted","Data":"72ebbb22ed48572155d88147af685c877ac8067937fd8c93811e851651d320bc"} Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.790549 4930 generic.go:334] "Generic (PLEG): container finished" podID="8f964d06-4f78-4e56-ac62-fc19a2020489" containerID="4f86d8a5b91c950152ea5feebeab50790467d609ad31c9a05c0b392b87c01c02" exitCode=0 Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.790589 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" event={"ID":"8f964d06-4f78-4e56-ac62-fc19a2020489","Type":"ContainerDied","Data":"4f86d8a5b91c950152ea5feebeab50790467d609ad31c9a05c0b392b87c01c02"} Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.793132 4930 generic.go:334] "Generic (PLEG): container finished" podID="79701ff1-8a11-4f24-9ab6-21f037545d5f" containerID="2bef53636a65aa4df31016d5e4a9dbca79a002bb165bc0e482564a6814cfcc43" exitCode=2 Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.793154 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"79701ff1-8a11-4f24-9ab6-21f037545d5f","Type":"ContainerDied","Data":"2bef53636a65aa4df31016d5e4a9dbca79a002bb165bc0e482564a6814cfcc43"} Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.828730 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" podUID="43277746-d9dc-4452-bcaf-8c4668ebf473" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.828766 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" podUID="561d468b-4138-4196-8652-38b9eb4519ea" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.828987 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" podUID="6c666666-f9ed-4183-9a92-e0334239aa3e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.121:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.828813 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-zx2fh container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.829195 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-zx2fh" podUID="bcd66746-45e7-4f01-a2ed-29da89486b3c" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.57:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.828834 4930 patch_prober.go:28] interesting pod/logging-loki-gateway-68bf8bdf7c-sm6c6 container/gateway namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.58:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.829254 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-gateway-68bf8bdf7c-sm6c6" podUID="51635511-c1d7-43e1-a692-bba69266bdf7" containerName="gateway" probeResult="failure" output="Get \"https://10.217.0.58:8081/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.829875 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.912604 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" podUID="046b75aa-f3a0-467d-8624-6a82aa748b32" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.122:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.912651 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" podUID="404966af-7f68-4667-b535-81136694bdb9" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.105:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: E0313 10:30:22.951032 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-4a1fbef37d7349b1be77763d642a29b52e831bc1c84f43d9151467d8e12a0db8.scope\": RecentStats: unable to find data in memory cache]" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.953581 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" podUID="7601eb51-d514-4a96-848c-af0eae32f54f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.953684 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.953748 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Readiness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": EOF" start-of-body= Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.953800 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": EOF" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.953883 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" podUID="7601eb51-d514-4a96-848c-af0eae32f54f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.953945 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.953932 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podUID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.124:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.954082 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.954198 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" podUID="c92ab5a7-1ad7-4584-9954-0047b879bc47" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.125:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.954510 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" podUID="c92ab5a7-1ad7-4584-9954-0047b879bc47" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.125:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.954574 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" podUID="43277746-d9dc-4452-bcaf-8c4668ebf473" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.126:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.954619 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" podUID="f338c47b-e688-4de8-8bd7-02be68a70ad3" containerName="webhook-server" probeResult="failure" output="Get \"http://10.217.0.97:7472/metrics\": dial tcp 10.217.0.97:7472: connect: connection refused" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.954823 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.954867 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.955636 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": EOF" start-of-body= Mar 13 10:30:22 crc kubenswrapper[4930]: I0313 10:30:22.955685 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": EOF" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.037610 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" podUID="2e04d1ac-077a-4dc5-91a4-c9b22c352ac2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.106:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.037704 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" podUID="4051662d-ae7a-4f38-9aa5-5eecab9693f3" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.108:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.079699 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Liveness probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.079819 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.079749 4930 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-zfz2k container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": dial tcp 10.217.0.78:8443: connect: connection refused" start-of-body= Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.079925 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" podUID="8f964d06-4f78-4e56-ac62-fc19a2020489" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": dial tcp 10.217.0.78:8443: connect: connection refused" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.079997 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" podUID="12a9ad60-3dff-443d-9f09-b57c4a373aa2" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.111:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.080159 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.080185 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.272600 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" podUID="4a2edceb-a605-4edd-8905-e35177c281c7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.110:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.397686 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" podUID="cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.96:8080/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.422385 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-68f45f9d9f-vsnn4" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.570646 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" podUID="47a82845-cf9f-4fc8-a35f-4548a9c7127e" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.114:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.652667 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" podUID="0d606b70-8e59-4303-963c-54c0be0a3800" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.116:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.653215 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" podUID="a28dc996-b000-4680-b9f1-d770e91f0cba" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.117:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.695694 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" podUID="585f7875-1138-4fcb-b89a-0a5f208f04b7" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.118:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.707413 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.717522 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/healthy\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.717603 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.717702 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/ready\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.720133 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="prometheus" containerStatusID={"Type":"cri-o","ID":"6eff41a2dd17dda4fc1c039e2a2403c0a42d01c5a48ac061aed74d6715199b83"} pod="openstack/prometheus-metric-storage-0" containerMessage="Container prometheus failed liveness probe, will be restarted" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.720223 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" containerID="cri-o://6eff41a2dd17dda4fc1c039e2a2403c0a42d01c5a48ac061aed74d6715199b83" gracePeriod=600 Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.731228 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podUID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.124:8081/readyz\": EOF" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.764399 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mt2n8" podUID="36c913ba-902f-480a-a1ba-ae8b95faa2d7" containerName="registry-server" probeResult="failure" output="command timed out" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.821084 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.829164 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.834078 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.834153 4930 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="4a1fbef37d7349b1be77763d642a29b52e831bc1c84f43d9151467d8e12a0db8" exitCode=1 Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.834272 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"4a1fbef37d7349b1be77763d642a29b52e831bc1c84f43d9151467d8e12a0db8"} Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.836796 4930 scope.go:117] "RemoveContainer" containerID="e3dc535836d240a06323d71572218900609f93fcf435b117c948297f522800cc" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.836982 4930 scope.go:117] "RemoveContainer" containerID="4a1fbef37d7349b1be77763d642a29b52e831bc1c84f43d9151467d8e12a0db8" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.843355 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_downloads-7954f5f757-22ldf_dc201d97-a254-45f7-8f60-f2c5dd726852/download-server/0.log" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.843515 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-22ldf" event={"ID":"dc201d97-a254-45f7-8f60-f2c5dd726852","Type":"ContainerStarted","Data":"80cd421601f8ce5d917b0fef001b3190b4e961ae885186eea74b792b9464a76c"} Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.843722 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.846352 4930 generic.go:334] "Generic (PLEG): container finished" podID="819e7015-1add-4a64-be13-d4ff9c73b228" containerID="168aa9a41e863b3e1cbb2cb4acc917874653648d080c859dd1cc0524ede1dedc" exitCode=0 Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.846409 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" event={"ID":"819e7015-1add-4a64-be13-d4ff9c73b228","Type":"ContainerDied","Data":"168aa9a41e863b3e1cbb2cb4acc917874653648d080c859dd1cc0524ede1dedc"} Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.871902 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" podUID="561d468b-4138-4196-8652-38b9eb4519ea" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.119:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.871983 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.872032 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:23 crc kubenswrapper[4930]: I0313 10:30:23.996676 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" podUID="7601eb51-d514-4a96-848c-af0eae32f54f" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.123:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.049217 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" podUID="bbf3e078-1f66-4bcc-a742-0835afcb3e79" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": dial tcp 10.217.0.98:7572: connect: connection refused" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.104636 4930 patch_prober.go:28] interesting pod/thanos-querier-7c466dcfcb-k7psd container/kube-rbac-proxy-web namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.104740 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/thanos-querier-7c466dcfcb-k7psd" podUID="951e59b6-d39a-4ec2-beda-4b6b8b3e20c8" containerName="kube-rbac-proxy-web" probeResult="failure" output="Get \"https://10.217.0.83:9091/-/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.121351 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.136907 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.287656 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" podUID="984e8970-7f0d-4ce8-87c2-56c787857225" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.127:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.377417 4930 patch_prober.go:28] interesting pod/controller-manager-7984f94f96-clbzb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.377524 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" podUID="73d513ce-8d01-4a5e-ad2c-a3f86c80668a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.452971 4930 patch_prober.go:28] interesting pod/route-controller-manager-55d94fcf7b-v9cp9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" start-of-body= Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.453348 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" podUID="23510207-60fc-496b-b907-224360e7e887" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.869467 4930 generic.go:334] "Generic (PLEG): container finished" podID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" containerID="8d450e36e2a6be6462e86bc3137c2d91e9b9dd1006f05f4fb76254cf36f5c093" exitCode=1 Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.869544 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" event={"ID":"1a25b1bf-af64-4f8c-920c-c450ebc66b41","Type":"ContainerDied","Data":"8d450e36e2a6be6462e86bc3137c2d91e9b9dd1006f05f4fb76254cf36f5c093"} Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.870980 4930 scope.go:117] "RemoveContainer" containerID="8d450e36e2a6be6462e86bc3137c2d91e9b9dd1006f05f4fb76254cf36f5c093" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.871953 4930 generic.go:334] "Generic (PLEG): container finished" podID="d60fb09e-fa42-482c-9228-d89352573e75" containerID="d551980d1274b85a05cbf3ca612f7a73e8af7baef61f6c5b053db6dfdc5b6347" exitCode=0 Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.872020 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" event={"ID":"d60fb09e-fa42-482c-9228-d89352573e75","Type":"ContainerDied","Data":"d551980d1274b85a05cbf3ca612f7a73e8af7baef61f6c5b053db6dfdc5b6347"} Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.874036 4930 generic.go:334] "Generic (PLEG): container finished" podID="bbf3e078-1f66-4bcc-a742-0835afcb3e79" containerID="9fe44cdc98892fa6b9a768d0a23c5444c75a778e1a00875feae8c0c303a524ed" exitCode=0 Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.874098 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" event={"ID":"bbf3e078-1f66-4bcc-a742-0835afcb3e79","Type":"ContainerDied","Data":"9fe44cdc98892fa6b9a768d0a23c5444c75a778e1a00875feae8c0c303a524ed"} Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.882140 4930 generic.go:334] "Generic (PLEG): container finished" podID="1e87ea83-860c-469b-93ba-1f9cbc7db958" containerID="da0e90531521761cb303522a5a41ba4b7c379d88504e0716cbb1fa3ec99fdc76" exitCode=143 Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.882248 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerDied","Data":"da0e90531521761cb303522a5a41ba4b7c379d88504e0716cbb1fa3ec99fdc76"} Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.887396 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" event={"ID":"143f048a-f720-4d97-92ac-c9766a7e39ab","Type":"ContainerStarted","Data":"a37ee81bdb7657f6bd31071e8155be88e218c77b9267493c18c84e7b43b9eb0a"} Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.887884 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.888167 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.888219 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.891149 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console-operator_console-operator-58897d9998-w957v_a2feeb9b-9255-4660-829f-23c296a2a1c3/console-operator/0.log" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.891480 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-w957v" event={"ID":"a2feeb9b-9255-4660-829f-23c296a2a1c3","Type":"ContainerStarted","Data":"5da5126ad4c5803ccb8f6511c4e86942f3437766984848e3d3788c279e0e1431"} Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.891679 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.891998 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.892064 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.894726 4930 generic.go:334] "Generic (PLEG): container finished" podID="1b960305-f6c1-4892-bbc2-70cce9c21546" containerID="7884531756cb3156917f38dc87d7796a7de383a57b50e29e1f711951f47dea05" exitCode=0 Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.894926 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" event={"ID":"1b960305-f6c1-4892-bbc2-70cce9c21546","Type":"ContainerDied","Data":"7884531756cb3156917f38dc87d7796a7de383a57b50e29e1f711951f47dea05"} Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.895809 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.895855 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:24 crc kubenswrapper[4930]: I0313 10:30:24.922390 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-zrwjj" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.067872 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/speaker-hc6zf" podUID="7f1370aa-666b-439e-9411-ec5e70360091" containerName="speaker" probeResult="failure" output="Get \"http://localhost:29150/metrics\": dial tcp [::1]:29150: connect: connection refused" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.426773 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/kube-state-metrics-0" podUID="79701ff1-8a11-4f24-9ab6-21f037545d5f" containerName="kube-state-metrics" probeResult="failure" output="Get \"https://10.217.1.23:8081/readyz\": dial tcp 10.217.1.23:8081: connect: connection refused" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.761978 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" probeResult="failure" output="command timed out" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.763394 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:30:25 crc kubenswrapper[4930]: E0313 10:30:25.765541 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1687564f2e703b32a639e036942f5c7161b2659f59f0f165489b992552ee549a" cmd=["sh","-c","if [ -x \"$(command -v curl)\" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x \"$(command -v wget)\" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi"] Mar 13 10:30:25 crc kubenswrapper[4930]: E0313 10:30:25.768768 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1687564f2e703b32a639e036942f5c7161b2659f59f0f165489b992552ee549a" cmd=["sh","-c","if [ -x \"$(command -v curl)\" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x \"$(command -v wget)\" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi"] Mar 13 10:30:25 crc kubenswrapper[4930]: E0313 10:30:25.771988 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1687564f2e703b32a639e036942f5c7161b2659f59f0f165489b992552ee549a" cmd=["sh","-c","if [ -x \"$(command -v curl)\" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x \"$(command -v wget)\" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi"] Mar 13 10:30:25 crc kubenswrapper[4930]: E0313 10:30:25.772051 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.785231 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.823275 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.829000 4930 patch_prober.go:28] interesting pod/nmstate-webhook-5f558f5558-z6585 container/nmstate-webhook namespace/openshift-nmstate: Readiness probe status=failure output="Get \"https://10.217.0.90:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.829058 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" podUID="9fc4b697-3810-4744-8c3f-2285ee8399e8" containerName="nmstate-webhook" probeResult="failure" output="Get \"https://10.217.0.90:9443/readyz\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.834933 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.912684 4930 patch_prober.go:28] interesting pod/loki-operator-controller-manager-7f7d54c9b6-mhwwb container/manager namespace/openshift-operators-redhat: Readiness probe status=failure output="Get \"http://10.217.0.50:8081/readyz\": dial tcp 10.217.0.50:8081: connect: connection refused" start-of-body= Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.913027 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" podUID="c17ab33f-40db-4679-8e37-e3c4d013ba29" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.50:8081/readyz\": dial tcp 10.217.0.50:8081: connect: connection refused" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.912685 4930 patch_prober.go:28] interesting pod/loki-operator-controller-manager-7f7d54c9b6-mhwwb container/manager namespace/openshift-operators-redhat: Liveness probe status=failure output="Get \"http://10.217.0.50:8081/healthz\": dial tcp 10.217.0.50:8081: connect: connection refused" start-of-body= Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.913092 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" podUID="c17ab33f-40db-4679-8e37-e3c4d013ba29" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.50:8081/healthz\": dial tcp 10.217.0.50:8081: connect: connection refused" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.933961 4930 generic.go:334] "Generic (PLEG): container finished" podID="08a4c763-07ef-4163-84ea-7a9bb39de0f1" containerID="e1607208fa9fe34d7a2cd486255d55261b51f641995f51b7b3fd83a2ff83ad9c" exitCode=0 Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.934098 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5rhcc" event={"ID":"08a4c763-07ef-4163-84ea-7a9bb39de0f1","Type":"ContainerDied","Data":"e1607208fa9fe34d7a2cd486255d55261b51f641995f51b7b3fd83a2ff83ad9c"} Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.945238 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hc6zf" event={"ID":"7f1370aa-666b-439e-9411-ec5e70360091","Type":"ContainerStarted","Data":"5c9019333d750937b20d6d0d8b3febd74a3549a32bf9f869fafa4b41753bfd53"} Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.945297 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hc6zf" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.950708 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"79701ff1-8a11-4f24-9ab6-21f037545d5f","Type":"ContainerStarted","Data":"ee53986ea18be7fc6991f477a5d0932abbaea3eece4cc95899c2a957da59e100"} Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.951067 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.954942 4930 generic.go:334] "Generic (PLEG): container finished" podID="d705c5d3-c140-479e-962c-958a2366753c" containerID="ef3feae5d76af1224b6c5c69d58de1290c3714a1add90314bbd94652c51a7da9" exitCode=0 Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.954992 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" event={"ID":"d705c5d3-c140-479e-962c-958a2366753c","Type":"ContainerDied","Data":"ef3feae5d76af1224b6c5c69d58de1290c3714a1add90314bbd94652c51a7da9"} Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.958887 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" event={"ID":"8f964d06-4f78-4e56-ac62-fc19a2020489","Type":"ContainerStarted","Data":"8db96ba661719fcad8814f88c8c6d46e6f7a477a2a141ebcbcc6063cdf15b423"} Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.959269 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.960358 4930 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-zfz2k container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": dial tcp 10.217.0.78:8443: connect: connection refused" start-of-body= Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.960500 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" podUID="8f964d06-4f78-4e56-ac62-fc19a2020489" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": dial tcp 10.217.0.78:8443: connect: connection refused" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.964801 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.967291 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.967381 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f264ce8447bd7cb99d07f9e0d7f5c5f77f8ec22eceeb83c50151cb43ca78710f"} Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.975306 4930 generic.go:334] "Generic (PLEG): container finished" podID="c17ab33f-40db-4679-8e37-e3c4d013ba29" containerID="78f7ff6f1df0496ae8d8dd1f0636778baed2566ac4daadc0df4d18d6c78e1eed" exitCode=1 Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.994746 4930 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8krq5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Mar 13 10:30:25 crc kubenswrapper[4930]: I0313 10:30:25.994877 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" podUID="d60fb09e-fa42-482c-9228-d89352573e75" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.002185 4930 generic.go:334] "Generic (PLEG): container finished" podID="47a82845-cf9f-4fc8-a35f-4548a9c7127e" containerID="b3f2db2d703fcbd1b9e8e356fa8567251136b01277698cdafa5e474fbfad9220" exitCode=1 Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.009542 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.009593 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.009693 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.009714 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.011375 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" event={"ID":"c17ab33f-40db-4679-8e37-e3c4d013ba29","Type":"ContainerDied","Data":"78f7ff6f1df0496ae8d8dd1f0636778baed2566ac4daadc0df4d18d6c78e1eed"} Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.011422 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" event={"ID":"d60fb09e-fa42-482c-9228-d89352573e75","Type":"ContainerStarted","Data":"55de283016ec55cf603abd7be266a0e9417fb515b2bc07054480f1f76b52c30f"} Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.011461 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.011474 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" event={"ID":"47a82845-cf9f-4fc8-a35f-4548a9c7127e","Type":"ContainerDied","Data":"b3f2db2d703fcbd1b9e8e356fa8567251136b01277698cdafa5e474fbfad9220"} Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.021378 4930 scope.go:117] "RemoveContainer" containerID="b3f2db2d703fcbd1b9e8e356fa8567251136b01277698cdafa5e474fbfad9220" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.021862 4930 scope.go:117] "RemoveContainer" containerID="78f7ff6f1df0496ae8d8dd1f0636778baed2566ac4daadc0df4d18d6c78e1eed" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.029629 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.029717 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.427311 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zptn9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" start-of-body= Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.427874 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.435967 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" podUID="d12dd258-eed7-41ff-b2e9-f124e189e9aa" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.120:8081/readyz\": dial tcp 10.217.0.120:8081: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.451335 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5995f4446f-mqvs5" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.659993 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]backend-http ok Mar 13 10:30:26 crc kubenswrapper[4930]: [+]has-synced ok Mar 13 10:30:26 crc kubenswrapper[4930]: [-]process-running failed: reason withheld Mar 13 10:30:26 crc kubenswrapper[4930]: healthz check failed Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.660269 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.672142 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.672264 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.672316 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.719846 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.719889 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.719954 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.720018 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.743177 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.743218 4930 patch_prober.go:28] interesting pod/console-operator-58897d9998-w957v container/console-operator namespace/openshift-console-operator: Liveness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.743248 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.743247 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-w957v" podUID="a2feeb9b-9255-4660-829f-23c296a2a1c3" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.762905 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/openstack-cell1-galera-0" podUID="3459cebe-06a7-428e-93ef-419677fbcb58" containerName="galera" probeResult="failure" output="command timed out" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.769648 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/heat-engine-f9949f6d4-jldnd" podUID="c8e68bf4-ac94-446c-883a-758ad1dfdb89" containerName="heat-engine" probeResult="failure" output="command timed out" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.769647 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-engine-f9949f6d4-jldnd" podUID="c8e68bf4-ac94-446c-883a-758ad1dfdb89" containerName="heat-engine" probeResult="failure" output="command timed out" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.787019 4930 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-j8gcf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.787051 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" podUID="1b960305-f6c1-4892-bbc2-70cce9c21546" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.812973 4930 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8krq5 container/packageserver namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.813027 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" podUID="d60fb09e-fa42-482c-9228-d89352573e75" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.812972 4930 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8krq5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Mar 13 10:30:26 crc kubenswrapper[4930]: I0313 10:30:26.813088 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" podUID="d60fb09e-fa42-482c-9228-d89352573e75" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.016407 4930 patch_prober.go:28] interesting pod/logging-loki-distributor-9c6b6d984-xfqch container/loki-distributor namespace/openshift-logging: Readiness probe status=failure output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.016475 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" podUID="e70aee0c-0c7e-428a-9fc7-c099088bd6a0" containerName="loki-distributor" probeResult="failure" output="Get \"https://10.217.0.54:3101/ready\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.043948 4930 generic.go:334] "Generic (PLEG): container finished" podID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerID="6eff41a2dd17dda4fc1c039e2a2403c0a42d01c5a48ac061aed74d6715199b83" exitCode=0 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.044079 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"54391e09-d1db-450f-9511-bfeaf8b0d817","Type":"ContainerDied","Data":"6eff41a2dd17dda4fc1c039e2a2403c0a42d01c5a48ac061aed74d6715199b83"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.047037 4930 generic.go:334] "Generic (PLEG): container finished" podID="23510207-60fc-496b-b907-224360e7e887" containerID="b5434c352bda7c745416492632729de7378d748dea3b07785cf57f04f8ffb95d" exitCode=0 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.047107 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" event={"ID":"23510207-60fc-496b-b907-224360e7e887","Type":"ContainerDied","Data":"b5434c352bda7c745416492632729de7378d748dea3b07785cf57f04f8ffb95d"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.056996 4930 generic.go:334] "Generic (PLEG): container finished" podID="d2ea88fb-18e6-4297-bb18-48d0e1d566be" containerID="636f02decec7b2fe9fec3d5b5cf8b8827855866bf3b417c9eb69ae8682673d24" exitCode=1 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.057044 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" event={"ID":"d2ea88fb-18e6-4297-bb18-48d0e1d566be","Type":"ContainerDied","Data":"636f02decec7b2fe9fec3d5b5cf8b8827855866bf3b417c9eb69ae8682673d24"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.057958 4930 scope.go:117] "RemoveContainer" containerID="636f02decec7b2fe9fec3d5b5cf8b8827855866bf3b417c9eb69ae8682673d24" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.067977 4930 generic.go:334] "Generic (PLEG): container finished" podID="15d7d52b-dc81-4400-bbce-353f00ff5103" containerID="f8eac0b1a2476de135d617d21cdcefaae2507867e2a9d881bb3deb2366d37e5b" exitCode=0 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.068079 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" event={"ID":"15d7d52b-dc81-4400-bbce-353f00ff5103","Type":"ContainerDied","Data":"f8eac0b1a2476de135d617d21cdcefaae2507867e2a9d881bb3deb2366d37e5b"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.072785 4930 generic.go:334] "Generic (PLEG): container finished" podID="d12dd258-eed7-41ff-b2e9-f124e189e9aa" containerID="75930369908ef1072aaf7fe4e5761c5ec63a3683818121a47bffb7adfd8d5149" exitCode=1 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.072883 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" event={"ID":"d12dd258-eed7-41ff-b2e9-f124e189e9aa","Type":"ContainerDied","Data":"75930369908ef1072aaf7fe4e5761c5ec63a3683818121a47bffb7adfd8d5149"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.074205 4930 scope.go:117] "RemoveContainer" containerID="75930369908ef1072aaf7fe4e5761c5ec63a3683818121a47bffb7adfd8d5149" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.077371 4930 generic.go:334] "Generic (PLEG): container finished" podID="d4640978-a913-419b-9faa-8b230f5e51f2" containerID="1cd5e3669981d79173656007193a66857ed29e6d11519fa9bc1873410a276c0f" exitCode=0 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.077421 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4640978-a913-419b-9faa-8b230f5e51f2","Type":"ContainerDied","Data":"1cd5e3669981d79173656007193a66857ed29e6d11519fa9bc1873410a276c0f"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.085635 4930 generic.go:334] "Generic (PLEG): container finished" podID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerID="9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959" exitCode=0 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.085738 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-568x4" event={"ID":"4dd00f07-858c-42be-a09e-746f709bbcdb","Type":"ContainerDied","Data":"9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.090527 4930 generic.go:334] "Generic (PLEG): container finished" podID="0d606b70-8e59-4303-963c-54c0be0a3800" containerID="6476e8dbcb813e652fee013be9726c717cf430303fba9ea91c144a0a4be6940d" exitCode=1 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.090615 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" event={"ID":"0d606b70-8e59-4303-963c-54c0be0a3800","Type":"ContainerDied","Data":"6476e8dbcb813e652fee013be9726c717cf430303fba9ea91c144a0a4be6940d"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.091747 4930 scope.go:117] "RemoveContainer" containerID="6476e8dbcb813e652fee013be9726c717cf430303fba9ea91c144a0a4be6940d" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.097955 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" event={"ID":"1a25b1bf-af64-4f8c-920c-c450ebc66b41","Type":"ContainerStarted","Data":"15ce15c08d845f46a5bbd5a2a3724e90f8c6706035494b1410e7d8ea4ccf6f0f"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.099449 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.121920 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-6dcbdf8bb8-kg8m5" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.121954 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.122000 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.122178 4930 generic.go:334] "Generic (PLEG): container finished" podID="2e04d1ac-077a-4dc5-91a4-c9b22c352ac2" containerID="16812e88101fbc43011a0e3cc1bddc4032f3210f3674087f01d47035ff1ca0f3" exitCode=1 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.122239 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" event={"ID":"2e04d1ac-077a-4dc5-91a4-c9b22c352ac2","Type":"ContainerDied","Data":"16812e88101fbc43011a0e3cc1bddc4032f3210f3674087f01d47035ff1ca0f3"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.123274 4930 scope.go:117] "RemoveContainer" containerID="16812e88101fbc43011a0e3cc1bddc4032f3210f3674087f01d47035ff1ca0f3" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.123831 4930 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-hrnlx container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" start-of-body= Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.123866 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" podUID="143f048a-f720-4d97-92ac-c9766a7e39ab" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.26:8443/healthz\": dial tcp 10.217.0.26:8443: connect: connection refused" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.131246 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-ff66c4dc9-l2bss" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.135887 4930 generic.go:334] "Generic (PLEG): container finished" podID="a28dc996-b000-4680-b9f1-d770e91f0cba" containerID="8fe24ba25156609c8f02fdd769b945deaf2a3c827ce6ef3473034cf5411efc6a" exitCode=1 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.135959 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" event={"ID":"a28dc996-b000-4680-b9f1-d770e91f0cba","Type":"ContainerDied","Data":"8fe24ba25156609c8f02fdd769b945deaf2a3c827ce6ef3473034cf5411efc6a"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.138118 4930 scope.go:117] "RemoveContainer" containerID="8fe24ba25156609c8f02fdd769b945deaf2a3c827ce6ef3473034cf5411efc6a" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.140292 4930 generic.go:334] "Generic (PLEG): container finished" podID="fd7bda25-ce82-4cc8-bc80-ecd97b9622cc" containerID="b7e3725e6882cd818f7397953267d9578e6b05b941ba8d004a00cfb1e03c0e93" exitCode=1 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.140405 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" event={"ID":"fd7bda25-ce82-4cc8-bc80-ecd97b9622cc","Type":"ContainerDied","Data":"b7e3725e6882cd818f7397953267d9578e6b05b941ba8d004a00cfb1e03c0e93"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.141488 4930 scope.go:117] "RemoveContainer" containerID="b7e3725e6882cd818f7397953267d9578e6b05b941ba8d004a00cfb1e03c0e93" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.147251 4930 generic.go:334] "Generic (PLEG): container finished" podID="c92ab5a7-1ad7-4584-9954-0047b879bc47" containerID="bef4938f9775237068dc2648d30bf2194f0e292f76ab09dd660e9ff1b6f3dfe7" exitCode=1 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.147345 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" event={"ID":"c92ab5a7-1ad7-4584-9954-0047b879bc47","Type":"ContainerDied","Data":"bef4938f9775237068dc2648d30bf2194f0e292f76ab09dd660e9ff1b6f3dfe7"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.148401 4930 scope.go:117] "RemoveContainer" containerID="bef4938f9775237068dc2648d30bf2194f0e292f76ab09dd660e9ff1b6f3dfe7" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.151032 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" event={"ID":"1b960305-f6c1-4892-bbc2-70cce9c21546","Type":"ContainerStarted","Data":"8444980ed8482dda07c97c7b4990d2cafc58bfde2a60dabd516307f2881e73c6"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.151399 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.151709 4930 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-j8gcf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.151743 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" podUID="1b960305-f6c1-4892-bbc2-70cce9c21546" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.160256 4930 generic.go:334] "Generic (PLEG): container finished" podID="73d513ce-8d01-4a5e-ad2c-a3f86c80668a" containerID="30bedad603a949a671f5e89c22fa86ae81a89a1bed0a3d726dda1b262d0cb66e" exitCode=0 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.160396 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" event={"ID":"73d513ce-8d01-4a5e-ad2c-a3f86c80668a","Type":"ContainerDied","Data":"30bedad603a949a671f5e89c22fa86ae81a89a1bed0a3d726dda1b262d0cb66e"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.234541 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-krmnx" event={"ID":"1e87ea83-860c-469b-93ba-1f9cbc7db958","Type":"ContainerStarted","Data":"37a16c7a7a7fade0a980522cac84d5cfd316de9efb4c929d74f96ca75f803f0e"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.263276 4930 generic.go:334] "Generic (PLEG): container finished" podID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerID="6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85" exitCode=0 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.263405 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kf9cp" event={"ID":"240f1f92-8c63-4c78-b7a9-52e9133f4617","Type":"ContainerDied","Data":"6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.270218 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" event={"ID":"819e7015-1add-4a64-be13-d4ff9c73b228","Type":"ContainerStarted","Data":"84396f273fbbd816d3aab1d356463208fde45f5492819b0af9c619a8488902e5"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.271292 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.271384 4930 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-mdgnp container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": dial tcp 10.217.0.13:8081: connect: connection refused" start-of-body= Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.271430 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": dial tcp 10.217.0.13:8081: connect: connection refused" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.276583 4930 generic.go:334] "Generic (PLEG): container finished" podID="cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a" containerID="1a6b7aecefef7fd19618182a43f445db9ec8c99bb1dc64ca39de9f5780dc2467" exitCode=1 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.276649 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" event={"ID":"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a","Type":"ContainerDied","Data":"1a6b7aecefef7fd19618182a43f445db9ec8c99bb1dc64ca39de9f5780dc2467"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.277561 4930 scope.go:117] "RemoveContainer" containerID="1a6b7aecefef7fd19618182a43f445db9ec8c99bb1dc64ca39de9f5780dc2467" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.289228 4930 generic.go:334] "Generic (PLEG): container finished" podID="046b75aa-f3a0-467d-8624-6a82aa748b32" containerID="3dda63aa85c8aed23d3f18135a76880272ed2f16e9b5fa1aa7705eb5b0fcf843" exitCode=1 Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.289301 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" event={"ID":"046b75aa-f3a0-467d-8624-6a82aa748b32","Type":"ContainerDied","Data":"3dda63aa85c8aed23d3f18135a76880272ed2f16e9b5fa1aa7705eb5b0fcf843"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.289954 4930 scope.go:117] "RemoveContainer" containerID="3dda63aa85c8aed23d3f18135a76880272ed2f16e9b5fa1aa7705eb5b0fcf843" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.314402 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" event={"ID":"bbf3e078-1f66-4bcc-a742-0835afcb3e79","Type":"ContainerStarted","Data":"9f7a858d553a3b2870c6518a2523da64a499e82dc8f7a09ac4fe2cdb747d94c1"} Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.314958 4930 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-8krq5 container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" start-of-body= Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.315027 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" podUID="d60fb09e-fa42-482c-9228-d89352573e75" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.39:5443/healthz\": dial tcp 10.217.0.39:5443: connect: connection refused" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.316064 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.318759 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" podUID="bbf3e078-1f66-4bcc-a742-0835afcb3e79" containerName="frr-k8s-webhook-server" probeResult="failure" output="Get \"http://10.217.0.98:7572/metrics\": dial tcp 10.217.0.98:7572: connect: connection refused" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.320495 4930 patch_prober.go:28] interesting pod/prometheus-operator-admission-webhook-f54c54754-zfz2k container/prometheus-operator-admission-webhook namespace/openshift-monitoring: Readiness probe status=failure output="Get \"https://10.217.0.78:8443/healthz\": dial tcp 10.217.0.78:8443: connect: connection refused" start-of-body= Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.320556 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" podUID="8f964d06-4f78-4e56-ac62-fc19a2020489" containerName="prometheus-operator-admission-webhook" probeResult="failure" output="Get \"https://10.217.0.78:8443/healthz\": dial tcp 10.217.0.78:8443: connect: connection refused" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.327367 4930 trace.go:236] Trace[1202289359]: "Calculate volume metrics of swift for pod openstack/swift-storage-0" (13-Mar-2026 10:30:21.333) (total time: 5993ms): Mar 13 10:30:27 crc kubenswrapper[4930]: Trace[1202289359]: [5.99373464s] [5.99373464s] END Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.329077 4930 trace.go:236] Trace[2142329099]: "Calculate volume metrics of glance for pod openstack/glance-default-external-api-0" (13-Mar-2026 10:30:19.963) (total time: 7342ms): Mar 13 10:30:27 crc kubenswrapper[4930]: Trace[2142329099]: [7.342855488s] [7.342855488s] END Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.337127 4930 trace.go:236] Trace[1292974642]: "Calculate volume metrics of persistence for pod openstack/rabbitmq-server-2" (13-Mar-2026 10:30:19.043) (total time: 8267ms): Mar 13 10:30:27 crc kubenswrapper[4930]: Trace[1292974642]: [8.267684361s] [8.267684361s] END Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.348232 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.411072 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/monitoring-plugin-7f784d5d74-hnt2f" Mar 13 10:30:27 crc kubenswrapper[4930]: E0313 10:30:27.421316 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1687564f2e703b32a639e036942f5c7161b2659f59f0f165489b992552ee549a" cmd=["sh","-c","if [ -x \"$(command -v curl)\" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x \"$(command -v wget)\" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi"] Mar 13 10:30:27 crc kubenswrapper[4930]: E0313 10:30:27.428703 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1687564f2e703b32a639e036942f5c7161b2659f59f0f165489b992552ee549a" cmd=["sh","-c","if [ -x \"$(command -v curl)\" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x \"$(command -v wget)\" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi"] Mar 13 10:30:27 crc kubenswrapper[4930]: E0313 10:30:27.438244 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1687564f2e703b32a639e036942f5c7161b2659f59f0f165489b992552ee549a" cmd=["sh","-c","if [ -x \"$(command -v curl)\" ]; then exec curl --fail http://localhost:9090/-/ready; elif [ -x \"$(command -v wget)\" ]; then exec wget -q -O /dev/null http://localhost:9090/-/ready; else exit 1; fi"] Mar 13 10:30:27 crc kubenswrapper[4930]: E0313 10:30:27.438332 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openshift-monitoring/prometheus-k8s-0" podUID="5f969789-8156-442b-b177-685669da5416" containerName="prometheus" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.510085 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Mar 13 10:30:27 crc kubenswrapper[4930]: I0313 10:30:27.767773 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-notification-agent" probeResult="failure" output="command timed out" Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.242267 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-cell1-galera-0" podUID="3459cebe-06a7-428e-93ef-419677fbcb58" containerName="galera" containerID="cri-o://b85c79faf68f9744648c98621a97e4ae9f30c296fa782f54ee36a56172c622e1" gracePeriod=19 Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.392538 4930 generic.go:334] "Generic (PLEG): container finished" podID="6c666666-f9ed-4183-9a92-e0334239aa3e" containerID="dfce868c7dce10d78c2a69dc4ef53eff1b3a6d17a6918af786fa4f7f7a2fc4a5" exitCode=1 Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.443450 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/openstack-galera-0" podUID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerName="galera" containerID="cri-o://c3c777a41680fafe3cd112f765e67426a68e7274d8368d2cff35e95608324d32" gracePeriod=18 Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.455768 4930 generic.go:334] "Generic (PLEG): container finished" podID="5f969789-8156-442b-b177-685669da5416" containerID="1687564f2e703b32a639e036942f5c7161b2659f59f0f165489b992552ee549a" exitCode=0 Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.475883 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovs-vswitchd" probeResult="failure" output=< Mar 13 10:30:28 crc kubenswrapper[4930]: 2026-03-13T10:30:28Z|00001|unixctl|WARN|failed to connect to /var/run/openvswitch/ovs-vswitchd.11.ctl Mar 13 10:30:28 crc kubenswrapper[4930]: ovs-appctl: cannot connect to "/var/run/openvswitch/ovs-vswitchd.11.ctl" (No such file or directory) Mar 13 10:30:28 crc kubenswrapper[4930]: ERROR - Failed retrieving ofproto/list from ovs-vswitchd Mar 13 10:30:28 crc kubenswrapper[4930]: > Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.479517 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" event={"ID":"6c666666-f9ed-4183-9a92-e0334239aa3e","Type":"ContainerDied","Data":"dfce868c7dce10d78c2a69dc4ef53eff1b3a6d17a6918af786fa4f7f7a2fc4a5"} Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.479558 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5f969789-8156-442b-b177-685669da5416","Type":"ContainerDied","Data":"1687564f2e703b32a639e036942f5c7161b2659f59f0f165489b992552ee549a"} Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.484051 4930 scope.go:117] "RemoveContainer" containerID="dfce868c7dce10d78c2a69dc4ef53eff1b3a6d17a6918af786fa4f7f7a2fc4a5" Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.485204 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-krmnx" Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.496983 4930 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="02a86f4198baffedc25ef8a9b11068c3e64ca1cddf968d18f4f2c26520c0db44" exitCode=0 Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.497203 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"02a86f4198baffedc25ef8a9b11068c3e64ca1cddf968d18f4f2c26520c0db44"} Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.531129 4930 generic.go:334] "Generic (PLEG): container finished" podID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerID="96e5faa3c3871a454979909e86e16277494094e8bfbd8123b70f92326e00fa79" exitCode=0 Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.531211 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" event={"ID":"afb89818-ec4f-4a94-8a7a-39e0e798616a","Type":"ContainerDied","Data":"96e5faa3c3871a454979909e86e16277494094e8bfbd8123b70f92326e00fa79"} Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.544977 4930 generic.go:334] "Generic (PLEG): container finished" podID="4051662d-ae7a-4f38-9aa5-5eecab9693f3" containerID="ed62b13cfd28c5f4a4a4fa9a940b2c319695c76af11cc94c3d55861434a47b97" exitCode=1 Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.545056 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" event={"ID":"4051662d-ae7a-4f38-9aa5-5eecab9693f3","Type":"ContainerDied","Data":"ed62b13cfd28c5f4a4a4fa9a940b2c319695c76af11cc94c3d55861434a47b97"} Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.545950 4930 scope.go:117] "RemoveContainer" containerID="ed62b13cfd28c5f4a4a4fa9a940b2c319695c76af11cc94c3d55861434a47b97" Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.554851 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5444994796-gc8fp_635a349e-a2f9-4fe3-b58a-72334ef95d6b/router/0.log" Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.554988 4930 generic.go:334] "Generic (PLEG): container finished" podID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerID="0d1c91ff9aa1863e77bdd6ace633558b349856411c7ecae8822b1abe4009f2b6" exitCode=137 Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.555090 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gc8fp" event={"ID":"635a349e-a2f9-4fe3-b58a-72334ef95d6b","Type":"ContainerDied","Data":"0d1c91ff9aa1863e77bdd6ace633558b349856411c7ecae8822b1abe4009f2b6"} Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.557872 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.557974 4930 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-j8gcf container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" start-of-body= Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.558003 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" podUID="1b960305-f6c1-4892-bbc2-70cce9c21546" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.25:8443/healthz\": dial tcp 10.217.0.25:8443: connect: connection refused" Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.558024 4930 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-mdgnp container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": dial tcp 10.217.0.13:8081: connect: connection refused" start-of-body= Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.558074 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": dial tcp 10.217.0.13:8081: connect: connection refused" Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.582982 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-krmnx" Mar 13 10:30:28 crc kubenswrapper[4930]: I0313 10:30:28.682422 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovs-vswitchd" containerID="cri-o://783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" gracePeriod=21 Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.020414 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.020758 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.027173 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.159695 4930 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-mdgnp container/operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": dial tcp 10.217.0.13:8081: connect: connection refused" start-of-body= Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.160033 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": dial tcp 10.217.0.13:8081: connect: connection refused" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.160274 4930 patch_prober.go:28] interesting pod/observability-operator-59bdc8b94-mdgnp container/operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.13:8081/healthz\": dial tcp 10.217.0.13:8081: connect: connection refused" start-of-body= Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.160302 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" podUID="819e7015-1add-4a64-be13-d4ff9c73b228" containerName="operator" probeResult="failure" output="Get \"http://10.217.0.13:8081/healthz\": dial tcp 10.217.0.13:8081: connect: connection refused" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.323849 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.329668 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-mbxlj" Mar 13 10:30:29 crc kubenswrapper[4930]: E0313 10:30:29.366577 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85 is running failed: container process not found" containerID="6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:29 crc kubenswrapper[4930]: E0313 10:30:29.374569 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85 is running failed: container process not found" containerID="6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:29 crc kubenswrapper[4930]: E0313 10:30:29.375112 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85 is running failed: container process not found" containerID="6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:29 crc kubenswrapper[4930]: E0313 10:30:29.375188 4930 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 6b08cca7bb20be765a426ef79060afcb65ccfb10e222b5419409619d9419cf85 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-kf9cp" podUID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerName="registry-server" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.591323 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"54391e09-d1db-450f-9511-bfeaf8b0d817","Type":"ContainerStarted","Data":"6c89c8a2cc524de7e8923222a4cfee42c0fde923fc378360bdf951785146c260"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.601236 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"0c3991716d2e740f754e8d5813acebde569758c79d765dd1a91e1155005bb767"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.601297 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.616885 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" event={"ID":"d705c5d3-c140-479e-962c-958a2366753c","Type":"ContainerStarted","Data":"14c5b12562236ee9ec53c387918a6099286105f9a4b5a9625035204ab9a88a8d"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.619339 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.619500 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zptn9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" start-of-body= Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.619535 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.644847 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" event={"ID":"c17ab33f-40db-4679-8e37-e3c4d013ba29","Type":"ContainerStarted","Data":"946da881a10e53cac5b8d2e15c99661c5124699b9f29936ac9704cac66b40a7b"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.645140 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.668313 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" event={"ID":"73d513ce-8d01-4a5e-ad2c-a3f86c80668a","Type":"ContainerStarted","Data":"366ee05f9a0a7b5dbab8f10cde925f38bbb3e72dd5d3d5258b1dffa94ca851b1"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.668368 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.668538 4930 patch_prober.go:28] interesting pod/controller-manager-7984f94f96-clbzb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.668587 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" podUID="73d513ce-8d01-4a5e-ad2c-a3f86c80668a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.688100 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" event={"ID":"d12dd258-eed7-41ff-b2e9-f124e189e9aa","Type":"ContainerStarted","Data":"1bde77748f4b800f0e601245ef74ba54c9e38e0ccd2f541c719c5f61b6ee40f9"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.689680 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.702779 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" event={"ID":"47a82845-cf9f-4fc8-a35f-4548a9c7127e","Type":"ContainerStarted","Data":"90a1dbcdb29192a8d35f5d58d5e04b7bd7e525bcbd09759707c436678185341f"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.703816 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.716547 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-5rhcc" event={"ID":"08a4c763-07ef-4163-84ea-7a9bb39de0f1","Type":"ContainerStarted","Data":"6e6511e42dea6176624939a14082d90c53940c002753b4f5e5c3bbe49c56dace"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.737154 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" event={"ID":"d2ea88fb-18e6-4297-bb18-48d0e1d566be","Type":"ContainerStarted","Data":"a793fb2ceaa823d91bfd5a7449850b06e350b6a48461dc29590f9e20da15a1b6"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.738489 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.744762 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" event={"ID":"cfac1c3a-6af2-46d2-a8aa-29c73b8b9b4a","Type":"ContainerStarted","Data":"7084a724fb8cb16a1b5f49f2df079684982baa8ea87d4523567852eb7d87591c"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.746135 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.757861 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" event={"ID":"2e04d1ac-077a-4dc5-91a4-c9b22c352ac2","Type":"ContainerStarted","Data":"e0397d9c1320476b5da9d129ab860b94fe27788a785797eeac3f14353920defd"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.758869 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.762640 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" event={"ID":"a28dc996-b000-4680-b9f1-d770e91f0cba","Type":"ContainerStarted","Data":"1e79a8274f549b6d5cda9d168dcc5888ce0c4318a5b639e543131d97c31e26f1"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.762928 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.772330 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/prometheus-k8s-0" event={"ID":"5f969789-8156-442b-b177-685669da5416","Type":"ContainerStarted","Data":"51742b5c990d7245e5ea0235000969f01dcd491455f2b358222f810ade6c2d9f"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.786644 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" event={"ID":"fd7bda25-ce82-4cc8-bc80-ecd97b9622cc","Type":"ContainerStarted","Data":"f568155ac775a231f2e937a4517d95f19b1897b86cc84b0305ffef0be5c26872"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.788118 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.836042 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" event={"ID":"0d606b70-8e59-4303-963c-54c0be0a3800","Type":"ContainerStarted","Data":"2610f8fdb771cb5d50d42e375859e9b1d4acdb947e6aa5868910e284458e5985"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.836409 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.849555 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" event={"ID":"23510207-60fc-496b-b907-224360e7e887","Type":"ContainerStarted","Data":"d69907a2a03c045fe5cf623df6ab7c55e973f7d41537dffc72bdea10a7a6a7ce"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.851503 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.851212 4930 patch_prober.go:28] interesting pod/route-controller-manager-55d94fcf7b-v9cp9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" start-of-body= Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.870288 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" podUID="23510207-60fc-496b-b907-224360e7e887" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.857309 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-9vpj8" event={"ID":"15d7d52b-dc81-4400-bbce-353f00ff5103","Type":"ContainerStarted","Data":"1387fbe39a52e7b0a6289a291c146bf64b70a7d712e5676086dc919c29e44ae0"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.882164 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" event={"ID":"c92ab5a7-1ad7-4584-9954-0047b879bc47","Type":"ContainerStarted","Data":"e2cb9f7320f2c6529becfc3396f9d32cc7a34c35a15fd002d8fb567261afb117"} Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.882736 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.917025 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 10:30:29 crc kubenswrapper[4930]: E0313 10:30:29.972557 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959 is running failed: container process not found" containerID="9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:29 crc kubenswrapper[4930]: E0313 10:30:29.973526 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959 is running failed: container process not found" containerID="9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:29 crc kubenswrapper[4930]: E0313 10:30:29.976383 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959 is running failed: container process not found" containerID="9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959" cmd=["grpc_health_probe","-addr=:50051"] Mar 13 10:30:29 crc kubenswrapper[4930]: E0313 10:30:29.976500 4930 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 9ab4040ce0fabf5601997944bf09c41bde0f403c981d9b50fdabc03cf8676959 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" Mar 13 10:30:29 crc kubenswrapper[4930]: I0313 10:30:29.987208 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mt2n8" podUID="36c913ba-902f-480a-a1ba-ae8b95faa2d7" containerName="registry-server" probeResult="failure" output=< Mar 13 10:30:29 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:30:29 crc kubenswrapper[4930]: > Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.314729 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-ff65d8fd4-h6zgj" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.415689 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.434168 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5964f64c48-6vcbx" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.457523 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-77b6666d85-npvw9" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.515673 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d9d6b584d-xlwng" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.717185 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.717578 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.717668 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/prometheus-metric-storage-0" podUID="54391e09-d1db-450f-9511-bfeaf8b0d817" containerName="prometheus" probeResult="failure" output="Get \"https://10.217.0.174:9090/-/ready\": dial tcp 10.217.0.174:9090: connect: connection refused" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.735955 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-569cc54c5-wwdw2" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.806267 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5f4f55cb5c-qx24r" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.907220 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ingress_router-default-5444994796-gc8fp_635a349e-a2f9-4fe3-b58a-72334ef95d6b/router/0.log" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.907311 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-gc8fp" event={"ID":"635a349e-a2f9-4fe3-b58a-72334ef95d6b","Type":"ContainerStarted","Data":"133a8358edcd63a840201e771230ac2bbd0fe2ade2c0cd2439a2b51fc9f0e255"} Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.927819 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" event={"ID":"046b75aa-f3a0-467d-8624-6a82aa748b32","Type":"ContainerStarted","Data":"112b6ff08c6ee2a4077c4e366067ff568252cc6a5c37cdcb0e00a1aa391378bd"} Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.929112 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.931808 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" event={"ID":"6c666666-f9ed-4183-9a92-e0334239aa3e","Type":"ContainerStarted","Data":"67d26138087e02f6c7f127b0471e485765311334a5ebf5d8959a17fff7fb7abd"} Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.934611 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.947264 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4640978-a913-419b-9faa-8b230f5e51f2","Type":"ContainerStarted","Data":"a7e1727f351e4f00388a3572c09ac857860d977688466021a0018423f9e40cc4"} Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.951899 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kf9cp" event={"ID":"240f1f92-8c63-4c78-b7a9-52e9133f4617","Type":"ContainerStarted","Data":"5662cfa1532d2d3874d7fd6f037dc28a7fac8427015e24204df8c223e2fa8d69"} Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.964668 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" event={"ID":"afb89818-ec4f-4a94-8a7a-39e0e798616a","Type":"ContainerStarted","Data":"3d036695f63e85be11d49fdeb5ef3a6789e6a15e78f0e96010cc9af2b3ceaf2c"} Mar 13 10:30:30 crc kubenswrapper[4930]: I0313 10:30:30.969115 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.037812 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-568x4" event={"ID":"4dd00f07-858c-42be-a09e-746f709bbcdb","Type":"ContainerStarted","Data":"94c0573e409a4d9db7cd885fd8bbdd199391d87ff286383d83562c0f362e8c38"} Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.045123 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" event={"ID":"4051662d-ae7a-4f38-9aa5-5eecab9693f3","Type":"ContainerStarted","Data":"e53a9ce5482d800c7b6b389ba03dd9a76ee8de771fd549e5bc8bd0a169c565ad"} Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.045745 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zptn9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" start-of-body= Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.045785 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.047060 4930 patch_prober.go:28] interesting pod/route-controller-manager-55d94fcf7b-v9cp9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" start-of-body= Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.047090 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" podUID="23510207-60fc-496b-b907-224360e7e887" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.047162 4930 patch_prober.go:28] interesting pod/controller-manager-7984f94f96-clbzb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.047205 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" podUID="73d513ce-8d01-4a5e-ad2c-a3f86c80668a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.194050 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-677c674df7-ntwzc" Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.232582 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" podUID="1a25b1bf-af64-4f8c-920c-c450ebc66b41" containerName="manager" probeResult="failure" output="Get \"http://10.217.0.124:8081/readyz\": dial tcp 10.217.0.124:8081: connect: connection refused" Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.356577 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6dd88c6f67-2jptm" Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.655667 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.656854 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 13 10:30:31 crc kubenswrapper[4930]: I0313 10:30:31.656903 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.021412 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.021467 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Liveness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.021499 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.021536 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.060713 4930 generic.go:334] "Generic (PLEG): container finished" podID="3459cebe-06a7-428e-93ef-419677fbcb58" containerID="b85c79faf68f9744648c98621a97e4ae9f30c296fa782f54ee36a56172c622e1" exitCode=0 Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.061057 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3459cebe-06a7-428e-93ef-419677fbcb58","Type":"ContainerDied","Data":"b85c79faf68f9744648c98621a97e4ae9f30c296fa782f54ee36a56172c622e1"} Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.061103 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"3459cebe-06a7-428e-93ef-419677fbcb58","Type":"ContainerStarted","Data":"93c4100014edbcf30356800defbf3f8c8e66c515e73ac42341f9f725b992ee18"} Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.062347 4930 patch_prober.go:28] interesting pod/route-controller-manager-55d94fcf7b-v9cp9 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" start-of-body= Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.062396 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" podUID="23510207-60fc-496b-b907-224360e7e887" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.68:8443/healthz\": dial tcp 10.217.0.68:8443: connect: connection refused" Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.062527 4930 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zptn9 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" start-of-body= Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.062545 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" podUID="d705c5d3-c140-479e-962c-958a2366753c" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.72:8080/healthz\": dial tcp 10.217.0.72:8080: connect: connection refused" Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.062549 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.062588 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.144999 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.145049 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.413610 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/prometheus-k8s-0" Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.510124 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-df66cd457-h9hb5" Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.660178 4930 patch_prober.go:28] interesting pod/router-default-5444994796-gc8fp container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 13 10:30:32 crc kubenswrapper[4930]: [-]has-synced failed: reason withheld Mar 13 10:30:32 crc kubenswrapper[4930]: [+]process-running ok Mar 13 10:30:32 crc kubenswrapper[4930]: healthz check failed Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.660243 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-gc8fp" podUID="635a349e-a2f9-4fe3-b58a-72334ef95d6b" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.690986 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:30:32 crc kubenswrapper[4930]: I0313 10:30:32.703039 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-operator-admission-webhook-f54c54754-zfz2k" Mar 13 10:30:33 crc kubenswrapper[4930]: I0313 10:30:33.082681 4930 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-d4ntr container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Mar 13 10:30:33 crc kubenswrapper[4930]: I0313 10:30:33.083012 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" podUID="afb89818-ec4f-4a94-8a7a-39e0e798616a" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/healthz\": dial tcp 10.217.0.19:8443: connect: connection refused" Mar 13 10:30:33 crc kubenswrapper[4930]: E0313 10:30:33.083561 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 10:30:33 crc kubenswrapper[4930]: E0313 10:30:33.098869 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 10:30:33 crc kubenswrapper[4930]: E0313 10:30:33.100358 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 10:30:33 crc kubenswrapper[4930]: E0313 10:30:33.100401 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovs-vswitchd" Mar 13 10:30:33 crc kubenswrapper[4930]: I0313 10:30:33.256212 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-6d5df8c674-hkqw8" Mar 13 10:30:33 crc kubenswrapper[4930]: I0313 10:30:33.376640 4930 patch_prober.go:28] interesting pod/controller-manager-7984f94f96-clbzb container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" start-of-body= Mar 13 10:30:33 crc kubenswrapper[4930]: I0313 10:30:33.376994 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" podUID="73d513ce-8d01-4a5e-ad2c-a3f86c80668a" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.67:8443/healthz\": dial tcp 10.217.0.67:8443: connect: connection refused" Mar 13 10:30:33 crc kubenswrapper[4930]: I0313 10:30:33.482873 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-krmnx" Mar 13 10:30:33 crc kubenswrapper[4930]: I0313 10:30:33.563340 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-9bhdc" Mar 13 10:30:33 crc kubenswrapper[4930]: I0313 10:30:33.658992 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 10:30:33 crc kubenswrapper[4930]: I0313 10:30:33.911008 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack-operators/openstack-operator-index-5rhcc" podUID="08a4c763-07ef-4163-84ea-7a9bb39de0f1" containerName="registry-server" probeResult="failure" output=< Mar 13 10:30:33 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:30:33 crc kubenswrapper[4930]: > Mar 13 10:30:34 crc kubenswrapper[4930]: I0313 10:30:34.173507 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 10:30:34 crc kubenswrapper[4930]: I0313 10:30:34.180718 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-gc8fp" Mar 13 10:30:34 crc kubenswrapper[4930]: E0313 10:30:34.319566 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c3c777a41680fafe3cd112f765e67426a68e7274d8368d2cff35e95608324d32" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 13 10:30:34 crc kubenswrapper[4930]: E0313 10:30:34.345809 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c3c777a41680fafe3cd112f765e67426a68e7274d8368d2cff35e95608324d32" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 13 10:30:34 crc kubenswrapper[4930]: E0313 10:30:34.347097 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="c3c777a41680fafe3cd112f765e67426a68e7274d8368d2cff35e95608324d32" cmd=["/bin/bash","/var/lib/operator-scripts/mysql_probe.sh","readiness"] Mar 13 10:30:34 crc kubenswrapper[4930]: E0313 10:30:34.347131 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/openstack-galera-0" podUID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerName="galera" Mar 13 10:30:34 crc kubenswrapper[4930]: I0313 10:30:34.461596 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-55d94fcf7b-v9cp9" Mar 13 10:30:34 crc kubenswrapper[4930]: I0313 10:30:34.789603 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-z6585" Mar 13 10:30:35 crc kubenswrapper[4930]: I0313 10:30:35.040766 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-d4ntr" Mar 13 10:30:35 crc kubenswrapper[4930]: I0313 10:30:35.079499 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-hc6zf" Mar 13 10:30:35 crc kubenswrapper[4930]: I0313 10:30:35.454628 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 13 10:30:35 crc kubenswrapper[4930]: I0313 10:30:35.562334 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 13 10:30:35 crc kubenswrapper[4930]: I0313 10:30:35.563644 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 13 10:30:35 crc kubenswrapper[4930]: I0313 10:30:35.627830 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:30:35 crc kubenswrapper[4930]: I0313 10:30:35.628174 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 10:30:35 crc kubenswrapper[4930]: I0313 10:30:35.629757 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cinder-scheduler" containerStatusID={"Type":"cri-o","ID":"7427f4f5722b28f8d487b62b99ea7f5be716124c600a214b0b1d7690a9e1c14d"} pod="openstack/cinder-scheduler-0" containerMessage="Container cinder-scheduler failed liveness probe, will be restarted" Mar 13 10:30:35 crc kubenswrapper[4930]: I0313 10:30:35.629826 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" containerID="cri-o://7427f4f5722b28f8d487b62b99ea7f5be716124c600a214b0b1d7690a9e1c14d" gracePeriod=30 Mar 13 10:30:35 crc kubenswrapper[4930]: I0313 10:30:35.947477 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-7f7d54c9b6-mhwwb" Mar 13 10:30:36 crc kubenswrapper[4930]: I0313 10:30:36.072940 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-9c6b6d984-xfqch" Mar 13 10:30:36 crc kubenswrapper[4930]: I0313 10:30:36.432212 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-zptn9" Mar 13 10:30:36 crc kubenswrapper[4930]: I0313 10:30:36.441591 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-557ccf57b7n5gzw" Mar 13 10:30:36 crc kubenswrapper[4930]: I0313 10:30:36.673233 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 13 10:30:36 crc kubenswrapper[4930]: I0313 10:30:36.673295 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 13 10:30:36 crc kubenswrapper[4930]: I0313 10:30:36.718976 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:36 crc kubenswrapper[4930]: I0313 10:30:36.719041 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:36 crc kubenswrapper[4930]: I0313 10:30:36.719102 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:36 crc kubenswrapper[4930]: I0313 10:30:36.719132 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:36 crc kubenswrapper[4930]: I0313 10:30:36.748650 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-w957v" Mar 13 10:30:36 crc kubenswrapper[4930]: I0313 10:30:36.795698 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-j8gcf" Mar 13 10:30:36 crc kubenswrapper[4930]: I0313 10:30:36.824232 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-8krq5" Mar 13 10:30:37 crc kubenswrapper[4930]: I0313 10:30:37.127718 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-hrnlx" Mar 13 10:30:37 crc kubenswrapper[4930]: I0313 10:30:37.214601 4930 generic.go:334] "Generic (PLEG): container finished" podID="8adee7e3-81a7-4dcf-ad19-d50e1900d3e0" containerID="c3c777a41680fafe3cd112f765e67426a68e7274d8368d2cff35e95608324d32" exitCode=0 Mar 13 10:30:37 crc kubenswrapper[4930]: I0313 10:30:37.214656 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0","Type":"ContainerDied","Data":"c3c777a41680fafe3cd112f765e67426a68e7274d8368d2cff35e95608324d32"} Mar 13 10:30:38 crc kubenswrapper[4930]: E0313 10:30:38.074667 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 10:30:38 crc kubenswrapper[4930]: E0313 10:30:38.080576 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 10:30:38 crc kubenswrapper[4930]: E0313 10:30:38.085327 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 10:30:38 crc kubenswrapper[4930]: E0313 10:30:38.085403 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovs-vswitchd" Mar 13 10:30:38 crc kubenswrapper[4930]: I0313 10:30:38.248785 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"8adee7e3-81a7-4dcf-ad19-d50e1900d3e0","Type":"ContainerStarted","Data":"4e77804045cf013a3ccdb6eb7a8c6941b22ddf050129ed80d05437be917775c3"} Mar 13 10:30:38 crc kubenswrapper[4930]: I0313 10:30:38.411808 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-7656bb8967-5vbb4" podUID="91a9c833-bbb1-4341-899b-d922b4066dc0" containerName="console" containerID="cri-o://f8d94b25364c5b095aa56613345a2b6e246aa3b132a9875afaede44395624bec" gracePeriod=12 Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.161413 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-mdgnp" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.365346 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.365480 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.416638 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jskr2"] Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.426386 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.571265 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jskr2"] Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.614555 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b25617-0c24-4567-9134-4f9783e97387-utilities\") pod \"certified-operators-jskr2\" (UID: \"72b25617-0c24-4567-9134-4f9783e97387\") " pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.614728 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ng8l\" (UniqueName: \"kubernetes.io/projected/72b25617-0c24-4567-9134-4f9783e97387-kube-api-access-8ng8l\") pod \"certified-operators-jskr2\" (UID: \"72b25617-0c24-4567-9134-4f9783e97387\") " pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.614791 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b25617-0c24-4567-9134-4f9783e97387-catalog-content\") pod \"certified-operators-jskr2\" (UID: \"72b25617-0c24-4567-9134-4f9783e97387\") " pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.718008 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b25617-0c24-4567-9134-4f9783e97387-utilities\") pod \"certified-operators-jskr2\" (UID: \"72b25617-0c24-4567-9134-4f9783e97387\") " pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.718076 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ng8l\" (UniqueName: \"kubernetes.io/projected/72b25617-0c24-4567-9134-4f9783e97387-kube-api-access-8ng8l\") pod \"certified-operators-jskr2\" (UID: \"72b25617-0c24-4567-9134-4f9783e97387\") " pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.718103 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b25617-0c24-4567-9134-4f9783e97387-catalog-content\") pod \"certified-operators-jskr2\" (UID: \"72b25617-0c24-4567-9134-4f9783e97387\") " pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.718669 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b25617-0c24-4567-9134-4f9783e97387-catalog-content\") pod \"certified-operators-jskr2\" (UID: \"72b25617-0c24-4567-9134-4f9783e97387\") " pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.719155 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b25617-0c24-4567-9134-4f9783e97387-utilities\") pod \"certified-operators-jskr2\" (UID: \"72b25617-0c24-4567-9134-4f9783e97387\") " pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.771835 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ng8l\" (UniqueName: \"kubernetes.io/projected/72b25617-0c24-4567-9134-4f9783e97387-kube-api-access-8ng8l\") pod \"certified-operators-jskr2\" (UID: \"72b25617-0c24-4567-9134-4f9783e97387\") " pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.783014 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.834269 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mt2n8" podUID="36c913ba-902f-480a-a1ba-ae8b95faa2d7" containerName="registry-server" probeResult="failure" output=< Mar 13 10:30:39 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:30:39 crc kubenswrapper[4930]: > Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.840220 4930 patch_prober.go:28] interesting pod/console-7656bb8967-5vbb4 container/console namespace/openshift-console: Readiness probe status=failure output="Get \"https://10.217.0.144:8443/health\": dial tcp 10.217.0.144:8443: connect: connection refused" start-of-body= Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.840569 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/console-7656bb8967-5vbb4" podUID="91a9c833-bbb1-4341-899b-d922b4066dc0" containerName="console" probeResult="failure" output="Get \"https://10.217.0.144:8443/health\": dial tcp 10.217.0.144:8443: connect: connection refused" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.971814 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-568x4" Mar 13 10:30:39 crc kubenswrapper[4930]: I0313 10:30:39.972350 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-568x4" Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.311658 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7656bb8967-5vbb4_91a9c833-bbb1-4341-899b-d922b4066dc0/console/0.log" Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.312053 4930 generic.go:334] "Generic (PLEG): container finished" podID="91a9c833-bbb1-4341-899b-d922b4066dc0" containerID="f8d94b25364c5b095aa56613345a2b6e246aa3b132a9875afaede44395624bec" exitCode=2 Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.312139 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7656bb8967-5vbb4" event={"ID":"91a9c833-bbb1-4341-899b-d922b4066dc0","Type":"ContainerDied","Data":"f8d94b25364c5b095aa56613345a2b6e246aa3b132a9875afaede44395624bec"} Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.338796 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-677bd678f7-kckp5" Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.347985 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-984cd4dcf-mcqf7" Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.400865 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.405115 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-66d56f6ff4-nsxgv" Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.438889 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-kf9cp" podUID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerName="registry-server" probeResult="failure" output=< Mar 13 10:30:40 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:30:40 crc kubenswrapper[4930]: > Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.594298 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6bbb499bbc-7xbzn" Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.634046 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-684f77d66d-rmm4c" Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.686917 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-658d4cdd5-tk79c" Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.712011 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-776c5696bf-545tf" Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.873363 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-bbc5b68f9-f45t2" Mar 13 10:30:40 crc kubenswrapper[4930]: I0313 10:30:40.901732 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-574d45c66c-zrr4m" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.042145 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6scqh"] Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.053260 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.075756 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6scqh"] Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.088493 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb475308-399c-4168-8e73-5816fb8c30e9-catalog-content\") pod \"redhat-marketplace-6scqh\" (UID: \"eb475308-399c-4168-8e73-5816fb8c30e9\") " pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.088598 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz8zg\" (UniqueName: \"kubernetes.io/projected/eb475308-399c-4168-8e73-5816fb8c30e9-kube-api-access-nz8zg\") pod \"redhat-marketplace-6scqh\" (UID: \"eb475308-399c-4168-8e73-5816fb8c30e9\") " pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.088840 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb475308-399c-4168-8e73-5816fb8c30e9-utilities\") pod \"redhat-marketplace-6scqh\" (UID: \"eb475308-399c-4168-8e73-5816fb8c30e9\") " pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.114670 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output=< Mar 13 10:30:41 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:30:41 crc kubenswrapper[4930]: > Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.191224 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb475308-399c-4168-8e73-5816fb8c30e9-utilities\") pod \"redhat-marketplace-6scqh\" (UID: \"eb475308-399c-4168-8e73-5816fb8c30e9\") " pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.191426 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb475308-399c-4168-8e73-5816fb8c30e9-catalog-content\") pod \"redhat-marketplace-6scqh\" (UID: \"eb475308-399c-4168-8e73-5816fb8c30e9\") " pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.191501 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz8zg\" (UniqueName: \"kubernetes.io/projected/eb475308-399c-4168-8e73-5816fb8c30e9-kube-api-access-nz8zg\") pod \"redhat-marketplace-6scqh\" (UID: \"eb475308-399c-4168-8e73-5816fb8c30e9\") " pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.192175 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb475308-399c-4168-8e73-5816fb8c30e9-utilities\") pod \"redhat-marketplace-6scqh\" (UID: \"eb475308-399c-4168-8e73-5816fb8c30e9\") " pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.212376 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz8zg\" (UniqueName: \"kubernetes.io/projected/eb475308-399c-4168-8e73-5816fb8c30e9-kube-api-access-nz8zg\") pod \"redhat-marketplace-6scqh\" (UID: \"eb475308-399c-4168-8e73-5816fb8c30e9\") " pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.227672 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-5ffc4556d7-9h5cs" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.234180 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb475308-399c-4168-8e73-5816fb8c30e9-catalog-content\") pod \"redhat-marketplace-6scqh\" (UID: \"eb475308-399c-4168-8e73-5816fb8c30e9\") " pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.281055 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-8xc7f" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.326194 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-7656bb8967-5vbb4_91a9c833-bbb1-4341-899b-d922b4066dc0/console/0.log" Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.326259 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-7656bb8967-5vbb4" event={"ID":"91a9c833-bbb1-4341-899b-d922b4066dc0","Type":"ContainerStarted","Data":"c2c425fdb82b024cd213481ab8b817c868f7335daf665ec23013577b769ccc69"} Mar 13 10:30:41 crc kubenswrapper[4930]: I0313 10:30:41.372599 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:42 crc kubenswrapper[4930]: I0313 10:30:42.338258 4930 generic.go:334] "Generic (PLEG): container finished" podID="3b093601-3b3c-4541-ba63-bccaeedc6606" containerID="b4090d891229754dfe821cb8773a39f79da55d7c41f0d185fcee1e308b3f0554" exitCode=1 Mar 13 10:30:42 crc kubenswrapper[4930]: I0313 10:30:42.338351 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3b093601-3b3c-4541-ba63-bccaeedc6606","Type":"ContainerDied","Data":"b4090d891229754dfe821cb8773a39f79da55d7c41f0d185fcee1e308b3f0554"} Mar 13 10:30:42 crc kubenswrapper[4930]: I0313 10:30:42.915616 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 10:30:42 crc kubenswrapper[4930]: I0313 10:30:42.956343 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-5rhcc" Mar 13 10:30:43 crc kubenswrapper[4930]: E0313 10:30:43.085705 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 10:30:43 crc kubenswrapper[4930]: E0313 10:30:43.088204 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 10:30:43 crc kubenswrapper[4930]: E0313 10:30:43.089973 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 10:30:43 crc kubenswrapper[4930]: E0313 10:30:43.090069 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovs-vswitchd" Mar 13 10:30:43 crc kubenswrapper[4930]: I0313 10:30:43.353139 4930 generic.go:334] "Generic (PLEG): container finished" podID="7365e082-8383-4b59-974c-dfdedb789f77" containerID="7427f4f5722b28f8d487b62b99ea7f5be716124c600a214b0b1d7690a9e1c14d" exitCode=0 Mar 13 10:30:43 crc kubenswrapper[4930]: I0313 10:30:43.353301 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7365e082-8383-4b59-974c-dfdedb789f77","Type":"ContainerDied","Data":"7427f4f5722b28f8d487b62b99ea7f5be716124c600a214b0b1d7690a9e1c14d"} Mar 13 10:30:43 crc kubenswrapper[4930]: I0313 10:30:43.384360 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7984f94f96-clbzb" Mar 13 10:30:43 crc kubenswrapper[4930]: I0313 10:30:43.536519 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6scqh"] Mar 13 10:30:43 crc kubenswrapper[4930]: I0313 10:30:43.622203 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jskr2"] Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.065485 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-2tktx" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.294922 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.295279 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.377850 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"3b093601-3b3c-4541-ba63-bccaeedc6606","Type":"ContainerDied","Data":"ffa0e3165df97abf8c31797782fe346109d8bfe39a7ece869adafae439631840"} Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.381542 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ffa0e3165df97abf8c31797782fe346109d8bfe39a7ece869adafae439631840" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.398680 4930 generic.go:334] "Generic (PLEG): container finished" podID="4f1e0b58-9978-423f-84bc-f974dc6347b4" containerID="8c960286a069fbf9bcd5f4c6337974f0f1c68a60d77a10e51960f73a826c9a7d" exitCode=137 Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.398760 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" event={"ID":"4f1e0b58-9978-423f-84bc-f974dc6347b4","Type":"ContainerDied","Data":"8c960286a069fbf9bcd5f4c6337974f0f1c68a60d77a10e51960f73a826c9a7d"} Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.410403 4930 generic.go:334] "Generic (PLEG): container finished" podID="eb475308-399c-4168-8e73-5816fb8c30e9" containerID="c7e7b0b13d783101548245c3d6eeb92023f88fdc0afe1d3745cbd5f7177b4725" exitCode=0 Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.411216 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6scqh" event={"ID":"eb475308-399c-4168-8e73-5816fb8c30e9","Type":"ContainerDied","Data":"c7e7b0b13d783101548245c3d6eeb92023f88fdc0afe1d3745cbd5f7177b4725"} Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.411259 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6scqh" event={"ID":"eb475308-399c-4168-8e73-5816fb8c30e9","Type":"ContainerStarted","Data":"01302c574110474b85e7f0931cf9a33ecb1191147a06b1ee311f82d17a53df5b"} Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.421697 4930 generic.go:334] "Generic (PLEG): container finished" podID="72b25617-0c24-4567-9134-4f9783e97387" containerID="fd7af99e4702111b803095fc9b71326c0cb2b2676d42d2dd0a4125e737600f5b" exitCode=0 Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.421742 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jskr2" event={"ID":"72b25617-0c24-4567-9134-4f9783e97387","Type":"ContainerDied","Data":"fd7af99e4702111b803095fc9b71326c0cb2b2676d42d2dd0a4125e737600f5b"} Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.421773 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jskr2" event={"ID":"72b25617-0c24-4567-9134-4f9783e97387","Type":"ContainerStarted","Data":"5fec651218e07599af9744e048a5122bf2d46adc98d1b8bbc1500c771d9e4f1e"} Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.426097 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.607527 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b093601-3b3c-4541-ba63-bccaeedc6606-config-data\") pod \"3b093601-3b3c-4541-ba63-bccaeedc6606\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.607647 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wpcf7\" (UniqueName: \"kubernetes.io/projected/3b093601-3b3c-4541-ba63-bccaeedc6606-kube-api-access-wpcf7\") pod \"3b093601-3b3c-4541-ba63-bccaeedc6606\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.607676 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"3b093601-3b3c-4541-ba63-bccaeedc6606\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.607713 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b093601-3b3c-4541-ba63-bccaeedc6606-test-operator-ephemeral-temporary\") pod \"3b093601-3b3c-4541-ba63-bccaeedc6606\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.607790 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b093601-3b3c-4541-ba63-bccaeedc6606-test-operator-ephemeral-workdir\") pod \"3b093601-3b3c-4541-ba63-bccaeedc6606\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.607877 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-ssh-key\") pod \"3b093601-3b3c-4541-ba63-bccaeedc6606\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.607942 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-openstack-config-secret\") pod \"3b093601-3b3c-4541-ba63-bccaeedc6606\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.607976 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3b093601-3b3c-4541-ba63-bccaeedc6606-openstack-config\") pod \"3b093601-3b3c-4541-ba63-bccaeedc6606\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.608020 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-ca-certs\") pod \"3b093601-3b3c-4541-ba63-bccaeedc6606\" (UID: \"3b093601-3b3c-4541-ba63-bccaeedc6606\") " Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.611375 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b093601-3b3c-4541-ba63-bccaeedc6606-config-data" (OuterVolumeSpecName: "config-data") pod "3b093601-3b3c-4541-ba63-bccaeedc6606" (UID: "3b093601-3b3c-4541-ba63-bccaeedc6606"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.614384 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b093601-3b3c-4541-ba63-bccaeedc6606-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "3b093601-3b3c-4541-ba63-bccaeedc6606" (UID: "3b093601-3b3c-4541-ba63-bccaeedc6606"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.614822 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b093601-3b3c-4541-ba63-bccaeedc6606-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "3b093601-3b3c-4541-ba63-bccaeedc6606" (UID: "3b093601-3b3c-4541-ba63-bccaeedc6606"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.673787 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b093601-3b3c-4541-ba63-bccaeedc6606-kube-api-access-wpcf7" (OuterVolumeSpecName: "kube-api-access-wpcf7") pod "3b093601-3b3c-4541-ba63-bccaeedc6606" (UID: "3b093601-3b3c-4541-ba63-bccaeedc6606"). InnerVolumeSpecName "kube-api-access-wpcf7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.680547 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "test-operator-logs") pod "3b093601-3b3c-4541-ba63-bccaeedc6606" (UID: "3b093601-3b3c-4541-ba63-bccaeedc6606"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.689556 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "3b093601-3b3c-4541-ba63-bccaeedc6606" (UID: "3b093601-3b3c-4541-ba63-bccaeedc6606"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.689666 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3b093601-3b3c-4541-ba63-bccaeedc6606" (UID: "3b093601-3b3c-4541-ba63-bccaeedc6606"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.692405 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "3b093601-3b3c-4541-ba63-bccaeedc6606" (UID: "3b093601-3b3c-4541-ba63-bccaeedc6606"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.722679 4930 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.722719 4930 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-ca-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.722728 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3b093601-3b3c-4541-ba63-bccaeedc6606-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.722740 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wpcf7\" (UniqueName: \"kubernetes.io/projected/3b093601-3b3c-4541-ba63-bccaeedc6606-kube-api-access-wpcf7\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.722778 4930 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.722788 4930 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3b093601-3b3c-4541-ba63-bccaeedc6606-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.722800 4930 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3b093601-3b3c-4541-ba63-bccaeedc6606-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.722811 4930 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3b093601-3b3c-4541-ba63-bccaeedc6606-ssh-key\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.753294 4930 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.821938 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b093601-3b3c-4541-ba63-bccaeedc6606-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "3b093601-3b3c-4541-ba63-bccaeedc6606" (UID: "3b093601-3b3c-4541-ba63-bccaeedc6606"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.829986 4930 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:44 crc kubenswrapper[4930]: I0313 10:30:44.830023 4930 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3b093601-3b3c-4541-ba63-bccaeedc6606-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 13 10:30:45 crc kubenswrapper[4930]: I0313 10:30:45.433847 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 13 10:30:45 crc kubenswrapper[4930]: I0313 10:30:45.735520 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Mar 13 10:30:45 crc kubenswrapper[4930]: I0313 10:30:45.742140 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.128353 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" podUID="37ebb02d-bc69-46d5-b9ac-3f676826de28" containerName="oauth-openshift" containerID="cri-o://34f191dfc2ed28b8f28030a7f3b536e07a821336bad5a53f8da29c19b66b5be9" gracePeriod=14 Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.458713 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6scqh" event={"ID":"eb475308-399c-4168-8e73-5816fb8c30e9","Type":"ContainerStarted","Data":"8689a8419f2ca8c4d4600a4c395c189c53e509edf1e59c056976d7231230ddf3"} Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.462545 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jskr2" event={"ID":"72b25617-0c24-4567-9134-4f9783e97387","Type":"ContainerStarted","Data":"d73f0e053d6b5ce889711fd5831e57cc9521cf198beb5dbefa12e2847b536c4d"} Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.468381 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" event={"ID":"4f1e0b58-9978-423f-84bc-f974dc6347b4","Type":"ContainerStarted","Data":"39c6f1f8cb03ca8cbd99ba462f3e264a4ac93d1a5508bf17f64dee5e6ea179bb"} Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.471225 4930 generic.go:334] "Generic (PLEG): container finished" podID="37ebb02d-bc69-46d5-b9ac-3f676826de28" containerID="34f191dfc2ed28b8f28030a7f3b536e07a821336bad5a53f8da29c19b66b5be9" exitCode=0 Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.471534 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" event={"ID":"37ebb02d-bc69-46d5-b9ac-3f676826de28","Type":"ContainerDied","Data":"34f191dfc2ed28b8f28030a7f3b536e07a821336bad5a53f8da29c19b66b5be9"} Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.673011 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.673081 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.673142 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.674238 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"f264ce8447bd7cb99d07f9e0d7f5c5f77f8ec22eceeb83c50151cb43ca78710f"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.674333 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://f264ce8447bd7cb99d07f9e0d7f5c5f77f8ec22eceeb83c50151cb43ca78710f" gracePeriod=30 Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.718717 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.718781 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.719087 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.719112 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.719143 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.719700 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.719777 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.720097 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="download-server" containerStatusID={"Type":"cri-o","ID":"80cd421601f8ce5d917b0fef001b3190b4e961ae885186eea74b792b9464a76c"} pod="openshift-console/downloads-7954f5f757-22ldf" containerMessage="Container download-server failed liveness probe, will be restarted" Mar 13 10:30:46 crc kubenswrapper[4930]: I0313 10:30:46.720139 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" containerID="cri-o://80cd421601f8ce5d917b0fef001b3190b4e961ae885186eea74b792b9464a76c" gracePeriod=2 Mar 13 10:30:47 crc kubenswrapper[4930]: I0313 10:30:47.500878 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" event={"ID":"37ebb02d-bc69-46d5-b9ac-3f676826de28","Type":"ContainerStarted","Data":"5e1c3c047ffd2a1bf70327afa512dd0e418943391da4a783b2685d47124f9617"} Mar 13 10:30:47 crc kubenswrapper[4930]: I0313 10:30:47.502100 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 10:30:47 crc kubenswrapper[4930]: I0313 10:30:47.501503 4930 patch_prober.go:28] interesting pod/oauth-openshift-69c75b6f68-wdg64 container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.66:6443/healthz\": dial tcp 10.217.0.66:6443: connect: connection refused" start-of-body= Mar 13 10:30:47 crc kubenswrapper[4930]: I0313 10:30:47.502345 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" podUID="37ebb02d-bc69-46d5-b9ac-3f676826de28" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.66:6443/healthz\": dial tcp 10.217.0.66:6443: connect: connection refused" Mar 13 10:30:47 crc kubenswrapper[4930]: I0313 10:30:47.504308 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"7365e082-8383-4b59-974c-dfdedb789f77","Type":"ContainerStarted","Data":"8e027fb98d7caab9b764611f44061b19a3c72c1158b4e4276cbda1588273dfdf"} Mar 13 10:30:47 crc kubenswrapper[4930]: I0313 10:30:47.507051 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_downloads-7954f5f757-22ldf_dc201d97-a254-45f7-8f60-f2c5dd726852/download-server/0.log" Mar 13 10:30:47 crc kubenswrapper[4930]: I0313 10:30:47.507096 4930 generic.go:334] "Generic (PLEG): container finished" podID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerID="80cd421601f8ce5d917b0fef001b3190b4e961ae885186eea74b792b9464a76c" exitCode=0 Mar 13 10:30:47 crc kubenswrapper[4930]: I0313 10:30:47.508118 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-22ldf" event={"ID":"dc201d97-a254-45f7-8f60-f2c5dd726852","Type":"ContainerDied","Data":"80cd421601f8ce5d917b0fef001b3190b4e961ae885186eea74b792b9464a76c"} Mar 13 10:30:47 crc kubenswrapper[4930]: I0313 10:30:47.508185 4930 scope.go:117] "RemoveContainer" containerID="401da0a18500414bd6d51283db6b432caf9b0add3d600d776c92ee41c49a1850" Mar 13 10:30:47 crc kubenswrapper[4930]: I0313 10:30:47.597148 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 13 10:30:48 crc kubenswrapper[4930]: E0313 10:30:48.064998 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 10:30:48 crc kubenswrapper[4930]: E0313 10:30:48.066280 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 10:30:48 crc kubenswrapper[4930]: E0313 10:30:48.067546 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" cmd=["/usr/local/bin/container-scripts/vswitchd_readiness.sh"] Mar 13 10:30:48 crc kubenswrapper[4930]: E0313 10:30:48.067596 4930 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovs-vswitchd" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.638739 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-69c75b6f68-wdg64" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.652824 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 13 10:30:48 crc kubenswrapper[4930]: E0313 10:30:48.653715 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b093601-3b3c-4541-ba63-bccaeedc6606" containerName="tempest-tests-tempest-tests-runner" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.653742 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b093601-3b3c-4541-ba63-bccaeedc6606" containerName="tempest-tests-tempest-tests-runner" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.654094 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b093601-3b3c-4541-ba63-bccaeedc6606" containerName="tempest-tests-tempest-tests-runner" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.655361 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.657541 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4srqg" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.668588 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.729940 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b94a38a5-0223-4e74-889a-95ac02363414\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.730194 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwkvw\" (UniqueName: \"kubernetes.io/projected/b94a38a5-0223-4e74-889a-95ac02363414-kube-api-access-dwkvw\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b94a38a5-0223-4e74-889a-95ac02363414\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.832600 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwkvw\" (UniqueName: \"kubernetes.io/projected/b94a38a5-0223-4e74-889a-95ac02363414-kube-api-access-dwkvw\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b94a38a5-0223-4e74-889a-95ac02363414\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.832729 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b94a38a5-0223-4e74-889a-95ac02363414\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.833316 4930 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b94a38a5-0223-4e74-889a-95ac02363414\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.862125 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwkvw\" (UniqueName: \"kubernetes.io/projected/b94a38a5-0223-4e74-889a-95ac02363414-kube-api-access-dwkvw\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b94a38a5-0223-4e74-889a-95ac02363414\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.868476 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"b94a38a5-0223-4e74-889a-95ac02363414\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 10:30:48 crc kubenswrapper[4930]: I0313 10:30:48.978100 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 13 10:30:49 crc kubenswrapper[4930]: I0313 10:30:49.560206 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-22ldf" event={"ID":"dc201d97-a254-45f7-8f60-f2c5dd726852","Type":"ContainerStarted","Data":"b960f62d3dfcbfa2d37993d7e07ddb4a213eba73b7bb0e095a89a5d491e38427"} Mar 13 10:30:49 crc kubenswrapper[4930]: I0313 10:30:49.560702 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 10:30:49 crc kubenswrapper[4930]: I0313 10:30:49.561046 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:49 crc kubenswrapper[4930]: I0313 10:30:49.561098 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:49 crc kubenswrapper[4930]: I0313 10:30:49.564477 4930 generic.go:334] "Generic (PLEG): container finished" podID="eb475308-399c-4168-8e73-5816fb8c30e9" containerID="8689a8419f2ca8c4d4600a4c395c189c53e509edf1e59c056976d7231230ddf3" exitCode=0 Mar 13 10:30:49 crc kubenswrapper[4930]: I0313 10:30:49.564531 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6scqh" event={"ID":"eb475308-399c-4168-8e73-5816fb8c30e9","Type":"ContainerDied","Data":"8689a8419f2ca8c4d4600a4c395c189c53e509edf1e59c056976d7231230ddf3"} Mar 13 10:30:49 crc kubenswrapper[4930]: I0313 10:30:49.674213 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 13 10:30:49 crc kubenswrapper[4930]: I0313 10:30:49.836431 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 10:30:49 crc kubenswrapper[4930]: I0313 10:30:49.836509 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 10:30:49 crc kubenswrapper[4930]: I0313 10:30:49.841322 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 10:30:49 crc kubenswrapper[4930]: I0313 10:30:49.907500 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mt2n8" podUID="36c913ba-902f-480a-a1ba-ae8b95faa2d7" containerName="registry-server" probeResult="failure" output=< Mar 13 10:30:49 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:30:49 crc kubenswrapper[4930]: > Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.419386 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-kf9cp" podUID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerName="registry-server" probeResult="failure" output=< Mar 13 10:30:50 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:30:50 crc kubenswrapper[4930]: > Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.578101 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6scqh" event={"ID":"eb475308-399c-4168-8e73-5816fb8c30e9","Type":"ContainerStarted","Data":"0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc"} Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.579896 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b94a38a5-0223-4e74-889a-95ac02363414","Type":"ContainerStarted","Data":"b061afdf872c0f19e1bd8dfd7b55768a2440e2f6c678bd0dda87654a9e938ba1"} Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.611284 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6scqh" podStartSLOduration=4.89183284 podStartE2EDuration="10.611264851s" podCreationTimestamp="2026-03-13 10:30:40 +0000 UTC" firstStartedPulling="2026-03-13 10:30:44.414331758 +0000 UTC m=+4685.164246435" lastFinishedPulling="2026-03-13 10:30:50.133763769 +0000 UTC m=+4690.883678446" observedRunningTime="2026-03-13 10:30:50.602597443 +0000 UTC m=+4691.352512120" watchObservedRunningTime="2026-03-13 10:30:50.611264851 +0000 UTC m=+4691.361179528" Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.639040 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zw9cp_20dc8601-3918-4d9c-993e-5ec14c0140b4/ovs-vswitchd/0.log" Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.640536 4930 generic.go:334] "Generic (PLEG): container finished" podID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerID="783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8" exitCode=137 Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.640561 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zw9cp" event={"ID":"20dc8601-3918-4d9c-993e-5ec14c0140b4","Type":"ContainerDied","Data":"783a889928f0949dd557acb88f5c0ed035ca9138f948a240bfcf6c923d4aa5e8"} Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.640626 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zw9cp" event={"ID":"20dc8601-3918-4d9c-993e-5ec14c0140b4","Type":"ContainerStarted","Data":"539c97744c464bf353f0103f58b6896abd3fca7c876640a0f0309251ff675a43"} Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.640650 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.641369 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="ovsdb-server" containerStatusID={"Type":"cri-o","ID":"3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba"} pod="openstack/ovn-controller-ovs-zw9cp" containerMessage="Container ovsdb-server failed liveness probe, will be restarted" Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.646722 4930 generic.go:334] "Generic (PLEG): container finished" podID="72b25617-0c24-4567-9134-4f9783e97387" containerID="d73f0e053d6b5ce889711fd5831e57cc9521cf198beb5dbefa12e2847b536c4d" exitCode=0 Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.646770 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jskr2" event={"ID":"72b25617-0c24-4567-9134-4f9783e97387","Type":"ContainerDied","Data":"d73f0e053d6b5ce889711fd5831e57cc9521cf198beb5dbefa12e2847b536c4d"} Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.649759 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.649831 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.653799 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-7656bb8967-5vbb4" Mar 13 10:30:50 crc kubenswrapper[4930]: E0313 10:30:50.804416 4930 handlers.go:78] "Exec lifecycle hook for Container in Pod failed" err=< Mar 13 10:30:50 crc kubenswrapper[4930]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Mar 13 10:30:50 crc kubenswrapper[4930]: + source /usr/local/bin/container-scripts/functions Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNBridge=br-int Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNRemote=tcp:localhost:6642 Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNEncapType=geneve Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNAvailabilityZones= Mar 13 10:30:50 crc kubenswrapper[4930]: ++ EnableChassisAsGateway=true Mar 13 10:30:50 crc kubenswrapper[4930]: ++ PhysicalNetworks= Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNHostName= Mar 13 10:30:50 crc kubenswrapper[4930]: ++ DB_FILE=/etc/openvswitch/conf.db Mar 13 10:30:50 crc kubenswrapper[4930]: ++ ovs_dir=/var/lib/openvswitch Mar 13 10:30:50 crc kubenswrapper[4930]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Mar 13 10:30:50 crc kubenswrapper[4930]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Mar 13 10:30:50 crc kubenswrapper[4930]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 10:30:50 crc kubenswrapper[4930]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 10:30:50 crc kubenswrapper[4930]: + cleanup_ovsdb_server_semaphore Mar 13 10:30:50 crc kubenswrapper[4930]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 10:30:50 crc kubenswrapper[4930]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Mar 13 10:30:50 crc kubenswrapper[4930]: > execCommand=["/usr/local/bin/container-scripts/stop-ovsdb-server.sh"] containerName="ovsdb-server" pod="openstack/ovn-controller-ovs-zw9cp" message=< Mar 13 10:30:50 crc kubenswrapper[4930]: Exiting ovsdb-server (5) [ OK ] Mar 13 10:30:50 crc kubenswrapper[4930]: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Mar 13 10:30:50 crc kubenswrapper[4930]: + source /usr/local/bin/container-scripts/functions Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNBridge=br-int Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNRemote=tcp:localhost:6642 Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNEncapType=geneve Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNAvailabilityZones= Mar 13 10:30:50 crc kubenswrapper[4930]: ++ EnableChassisAsGateway=true Mar 13 10:30:50 crc kubenswrapper[4930]: ++ PhysicalNetworks= Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNHostName= Mar 13 10:30:50 crc kubenswrapper[4930]: ++ DB_FILE=/etc/openvswitch/conf.db Mar 13 10:30:50 crc kubenswrapper[4930]: ++ ovs_dir=/var/lib/openvswitch Mar 13 10:30:50 crc kubenswrapper[4930]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Mar 13 10:30:50 crc kubenswrapper[4930]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Mar 13 10:30:50 crc kubenswrapper[4930]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 10:30:50 crc kubenswrapper[4930]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 10:30:50 crc kubenswrapper[4930]: + cleanup_ovsdb_server_semaphore Mar 13 10:30:50 crc kubenswrapper[4930]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 10:30:50 crc kubenswrapper[4930]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Mar 13 10:30:50 crc kubenswrapper[4930]: > Mar 13 10:30:50 crc kubenswrapper[4930]: E0313 10:30:50.806800 4930 kuberuntime_container.go:691] "PreStop hook failed" err=< Mar 13 10:30:50 crc kubenswrapper[4930]: command '/usr/local/bin/container-scripts/stop-ovsdb-server.sh' exited with 137: ++ dirname /usr/local/bin/container-scripts/stop-ovsdb-server.sh Mar 13 10:30:50 crc kubenswrapper[4930]: + source /usr/local/bin/container-scripts/functions Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNBridge=br-int Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNRemote=tcp:localhost:6642 Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNEncapType=geneve Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNAvailabilityZones= Mar 13 10:30:50 crc kubenswrapper[4930]: ++ EnableChassisAsGateway=true Mar 13 10:30:50 crc kubenswrapper[4930]: ++ PhysicalNetworks= Mar 13 10:30:50 crc kubenswrapper[4930]: ++ OVNHostName= Mar 13 10:30:50 crc kubenswrapper[4930]: ++ DB_FILE=/etc/openvswitch/conf.db Mar 13 10:30:50 crc kubenswrapper[4930]: ++ ovs_dir=/var/lib/openvswitch Mar 13 10:30:50 crc kubenswrapper[4930]: ++ FLOWS_RESTORE_SCRIPT=/var/lib/openvswitch/flows-script Mar 13 10:30:50 crc kubenswrapper[4930]: ++ FLOWS_RESTORE_DIR=/var/lib/openvswitch/saved-flows Mar 13 10:30:50 crc kubenswrapper[4930]: ++ SAFE_TO_STOP_OVSDB_SERVER_SEMAPHORE=/var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 10:30:50 crc kubenswrapper[4930]: + '[' '!' -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server ']' Mar 13 10:30:50 crc kubenswrapper[4930]: + cleanup_ovsdb_server_semaphore Mar 13 10:30:50 crc kubenswrapper[4930]: + rm -f /var/lib/openvswitch/is_safe_to_stop_ovsdb_server Mar 13 10:30:50 crc kubenswrapper[4930]: + /usr/share/openvswitch/scripts/ovs-ctl stop --no-ovs-vswitchd Mar 13 10:30:50 crc kubenswrapper[4930]: > pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovsdb-server" containerID="cri-o://3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba" Mar 13 10:30:50 crc kubenswrapper[4930]: I0313 10:30:50.806859 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovsdb-server" containerID="cri-o://3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba" gracePeriod=30 Mar 13 10:30:51 crc kubenswrapper[4930]: I0313 10:30:51.041044 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output=< Mar 13 10:30:51 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:30:51 crc kubenswrapper[4930]: > Mar 13 10:30:51 crc kubenswrapper[4930]: I0313 10:30:51.372894 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:51 crc kubenswrapper[4930]: I0313 10:30:51.373280 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:30:51 crc kubenswrapper[4930]: I0313 10:30:51.663299 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zw9cp_20dc8601-3918-4d9c-993e-5ec14c0140b4/ovs-vswitchd/0.log" Mar 13 10:30:51 crc kubenswrapper[4930]: I0313 10:30:51.665368 4930 generic.go:334] "Generic (PLEG): container finished" podID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerID="3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba" exitCode=0 Mar 13 10:30:51 crc kubenswrapper[4930]: I0313 10:30:51.665403 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zw9cp" event={"ID":"20dc8601-3918-4d9c-993e-5ec14c0140b4","Type":"ContainerDied","Data":"3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba"} Mar 13 10:30:52 crc kubenswrapper[4930]: I0313 10:30:52.439954 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-6scqh" podUID="eb475308-399c-4168-8e73-5816fb8c30e9" containerName="registry-server" probeResult="failure" output=< Mar 13 10:30:52 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:30:52 crc kubenswrapper[4930]: > Mar 13 10:30:52 crc kubenswrapper[4930]: I0313 10:30:52.629582 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:30:52 crc kubenswrapper[4930]: I0313 10:30:52.679301 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"b94a38a5-0223-4e74-889a-95ac02363414","Type":"ContainerStarted","Data":"5a2c7a5ea9163cb2d4e684949dbefa0525cc26f28123beb29363a2d5c07c1d4a"} Mar 13 10:30:52 crc kubenswrapper[4930]: I0313 10:30:52.682407 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-zw9cp_20dc8601-3918-4d9c-993e-5ec14c0140b4/ovs-vswitchd/0.log" Mar 13 10:30:52 crc kubenswrapper[4930]: I0313 10:30:52.683378 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-zw9cp" event={"ID":"20dc8601-3918-4d9c-993e-5ec14c0140b4","Type":"ContainerStarted","Data":"75c74fde897c437c4b48b71729a066c4058e5e7d3a61d8ed630e9e6d044e0051"} Mar 13 10:30:52 crc kubenswrapper[4930]: E0313 10:30:52.683797 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba is running failed: container process not found" containerID="3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 10:30:52 crc kubenswrapper[4930]: E0313 10:30:52.684033 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba is running failed: container process not found" containerID="3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 10:30:52 crc kubenswrapper[4930]: E0313 10:30:52.684244 4930 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba is running failed: container process not found" containerID="3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba" cmd=["/usr/local/bin/container-scripts/ovsdb_server_readiness.sh"] Mar 13 10:30:52 crc kubenswrapper[4930]: E0313 10:30:52.684279 4930 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3544e876dcbeb1e75550ba18960209b1d3592d78bbf203b19e3a33d21e43d6ba is running failed: container process not found" probeType="Readiness" pod="openstack/ovn-controller-ovs-zw9cp" podUID="20dc8601-3918-4d9c-993e-5ec14c0140b4" containerName="ovsdb-server" Mar 13 10:30:52 crc kubenswrapper[4930]: I0313 10:30:52.686491 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jskr2" event={"ID":"72b25617-0c24-4567-9134-4f9783e97387","Type":"ContainerStarted","Data":"d30bd19fdd7ac708a5d36391a1303d851c5945faf75716fa6af1aab0e3373d8d"} Mar 13 10:30:52 crc kubenswrapper[4930]: I0313 10:30:52.704782 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.780691163 podStartE2EDuration="4.704758031s" podCreationTimestamp="2026-03-13 10:30:48 +0000 UTC" firstStartedPulling="2026-03-13 10:30:49.688347035 +0000 UTC m=+4690.438261702" lastFinishedPulling="2026-03-13 10:30:51.612413893 +0000 UTC m=+4692.362328570" observedRunningTime="2026-03-13 10:30:52.701602902 +0000 UTC m=+4693.451517579" watchObservedRunningTime="2026-03-13 10:30:52.704758031 +0000 UTC m=+4693.454672728" Mar 13 10:30:52 crc kubenswrapper[4930]: I0313 10:30:52.751284 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jskr2" podStartSLOduration=6.239474904 podStartE2EDuration="13.751262431s" podCreationTimestamp="2026-03-13 10:30:39 +0000 UTC" firstStartedPulling="2026-03-13 10:30:44.424551335 +0000 UTC m=+4685.174466012" lastFinishedPulling="2026-03-13 10:30:51.936338852 +0000 UTC m=+4692.686253539" observedRunningTime="2026-03-13 10:30:52.74125388 +0000 UTC m=+4693.491168557" watchObservedRunningTime="2026-03-13 10:30:52.751262431 +0000 UTC m=+4693.501177108" Mar 13 10:30:53 crc kubenswrapper[4930]: I0313 10:30:53.069562 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 10:30:56 crc kubenswrapper[4930]: I0313 10:30:56.718817 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:56 crc kubenswrapper[4930]: I0313 10:30:56.718817 4930 patch_prober.go:28] interesting pod/downloads-7954f5f757-22ldf container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" start-of-body= Mar 13 10:30:56 crc kubenswrapper[4930]: I0313 10:30:56.719511 4930 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:56 crc kubenswrapper[4930]: I0313 10:30:56.719456 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-22ldf" podUID="dc201d97-a254-45f7-8f60-f2c5dd726852" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.22:8080/\": dial tcp 10.217.0.22:8080: connect: connection refused" Mar 13 10:30:57 crc kubenswrapper[4930]: I0313 10:30:57.619823 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:30:59 crc kubenswrapper[4930]: I0313 10:30:59.784390 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:59 crc kubenswrapper[4930]: I0313 10:30:59.785088 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:30:59 crc kubenswrapper[4930]: I0313 10:30:59.846632 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-mt2n8" podUID="36c913ba-902f-480a-a1ba-ae8b95faa2d7" containerName="registry-server" probeResult="failure" output=< Mar 13 10:30:59 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:30:59 crc kubenswrapper[4930]: > Mar 13 10:31:00 crc kubenswrapper[4930]: I0313 10:31:00.839066 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jskr2" podUID="72b25617-0c24-4567-9134-4f9783e97387" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:00 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:00 crc kubenswrapper[4930]: > Mar 13 10:31:00 crc kubenswrapper[4930]: I0313 10:31:00.923369 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-kf9cp" podUID="240f1f92-8c63-4c78-b7a9-52e9133f4617" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:00 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:00 crc kubenswrapper[4930]: > Mar 13 10:31:01 crc kubenswrapper[4930]: I0313 10:31:01.020619 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:01 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:01 crc kubenswrapper[4930]: > Mar 13 10:31:02 crc kubenswrapper[4930]: I0313 10:31:02.426404 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-6scqh" podUID="eb475308-399c-4168-8e73-5816fb8c30e9" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:02 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:02 crc kubenswrapper[4930]: > Mar 13 10:31:02 crc kubenswrapper[4930]: I0313 10:31:02.432223 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5d4f94cfb6-bd6n5" Mar 13 10:31:02 crc kubenswrapper[4930]: I0313 10:31:02.625019 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:31:06 crc kubenswrapper[4930]: I0313 10:31:06.727352 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-22ldf" Mar 13 10:31:07 crc kubenswrapper[4930]: I0313 10:31:07.616649 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:31:09 crc kubenswrapper[4930]: I0313 10:31:09.085986 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:31:09 crc kubenswrapper[4930]: I0313 10:31:09.151918 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mt2n8" Mar 13 10:31:09 crc kubenswrapper[4930]: I0313 10:31:09.399614 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mt2n8"] Mar 13 10:31:09 crc kubenswrapper[4930]: I0313 10:31:09.432925 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 10:31:09 crc kubenswrapper[4930]: I0313 10:31:09.492663 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kf9cp" Mar 13 10:31:09 crc kubenswrapper[4930]: I0313 10:31:09.749003 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g6hb8"] Mar 13 10:31:09 crc kubenswrapper[4930]: I0313 10:31:09.749629 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-g6hb8" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerName="registry-server" containerID="cri-o://da6c0c379d767f63fba70506d469a63802e501f36556a8df4f7036d5281ac4a1" gracePeriod=2 Mar 13 10:31:10 crc kubenswrapper[4930]: I0313 10:31:10.839701 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jskr2" podUID="72b25617-0c24-4567-9134-4f9783e97387" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:10 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:10 crc kubenswrapper[4930]: > Mar 13 10:31:10 crc kubenswrapper[4930]: I0313 10:31:10.983785 4930 generic.go:334] "Generic (PLEG): container finished" podID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerID="da6c0c379d767f63fba70506d469a63802e501f36556a8df4f7036d5281ac4a1" exitCode=0 Mar 13 10:31:10 crc kubenswrapper[4930]: I0313 10:31:10.984602 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g6hb8" event={"ID":"2b4f2c08-b04f-40e6-a90c-256225efe861","Type":"ContainerDied","Data":"da6c0c379d767f63fba70506d469a63802e501f36556a8df4f7036d5281ac4a1"} Mar 13 10:31:11 crc kubenswrapper[4930]: I0313 10:31:11.040456 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:11 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:11 crc kubenswrapper[4930]: > Mar 13 10:31:11 crc kubenswrapper[4930]: I0313 10:31:11.757543 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g6hb8" Mar 13 10:31:11 crc kubenswrapper[4930]: I0313 10:31:11.786630 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdpzp\" (UniqueName: \"kubernetes.io/projected/2b4f2c08-b04f-40e6-a90c-256225efe861-kube-api-access-zdpzp\") pod \"2b4f2c08-b04f-40e6-a90c-256225efe861\" (UID: \"2b4f2c08-b04f-40e6-a90c-256225efe861\") " Mar 13 10:31:11 crc kubenswrapper[4930]: I0313 10:31:11.786794 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b4f2c08-b04f-40e6-a90c-256225efe861-utilities\") pod \"2b4f2c08-b04f-40e6-a90c-256225efe861\" (UID: \"2b4f2c08-b04f-40e6-a90c-256225efe861\") " Mar 13 10:31:11 crc kubenswrapper[4930]: I0313 10:31:11.787058 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b4f2c08-b04f-40e6-a90c-256225efe861-catalog-content\") pod \"2b4f2c08-b04f-40e6-a90c-256225efe861\" (UID: \"2b4f2c08-b04f-40e6-a90c-256225efe861\") " Mar 13 10:31:11 crc kubenswrapper[4930]: I0313 10:31:11.788200 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b4f2c08-b04f-40e6-a90c-256225efe861-utilities" (OuterVolumeSpecName: "utilities") pod "2b4f2c08-b04f-40e6-a90c-256225efe861" (UID: "2b4f2c08-b04f-40e6-a90c-256225efe861"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:31:11 crc kubenswrapper[4930]: I0313 10:31:11.806265 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b4f2c08-b04f-40e6-a90c-256225efe861-kube-api-access-zdpzp" (OuterVolumeSpecName: "kube-api-access-zdpzp") pod "2b4f2c08-b04f-40e6-a90c-256225efe861" (UID: "2b4f2c08-b04f-40e6-a90c-256225efe861"). InnerVolumeSpecName "kube-api-access-zdpzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:11 crc kubenswrapper[4930]: I0313 10:31:11.868987 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b4f2c08-b04f-40e6-a90c-256225efe861-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b4f2c08-b04f-40e6-a90c-256225efe861" (UID: "2b4f2c08-b04f-40e6-a90c-256225efe861"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:31:11 crc kubenswrapper[4930]: I0313 10:31:11.890490 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b4f2c08-b04f-40e6-a90c-256225efe861-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:11 crc kubenswrapper[4930]: I0313 10:31:11.890530 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b4f2c08-b04f-40e6-a90c-256225efe861-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:11 crc kubenswrapper[4930]: I0313 10:31:11.890545 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdpzp\" (UniqueName: \"kubernetes.io/projected/2b4f2c08-b04f-40e6-a90c-256225efe861-kube-api-access-zdpzp\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:12 crc kubenswrapper[4930]: I0313 10:31:12.036373 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-g6hb8" event={"ID":"2b4f2c08-b04f-40e6-a90c-256225efe861","Type":"ContainerDied","Data":"d7da926ac2d9ba108fb69324c81c5b1a78c95f6a49da6eaeb895a0f55e033be5"} Mar 13 10:31:12 crc kubenswrapper[4930]: I0313 10:31:12.036496 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-g6hb8" Mar 13 10:31:12 crc kubenswrapper[4930]: I0313 10:31:12.037363 4930 scope.go:117] "RemoveContainer" containerID="da6c0c379d767f63fba70506d469a63802e501f36556a8df4f7036d5281ac4a1" Mar 13 10:31:12 crc kubenswrapper[4930]: I0313 10:31:12.085904 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-g6hb8"] Mar 13 10:31:12 crc kubenswrapper[4930]: I0313 10:31:12.100087 4930 scope.go:117] "RemoveContainer" containerID="220750210a0a051baa308d129f907fef78ef4362348c8d034a8333034eaadcfc" Mar 13 10:31:12 crc kubenswrapper[4930]: I0313 10:31:12.100339 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-g6hb8"] Mar 13 10:31:12 crc kubenswrapper[4930]: I0313 10:31:12.186028 4930 scope.go:117] "RemoveContainer" containerID="e9b75150c047afd99ffe8d248526ab230c1c325dd118d491cac47e43ea408ec2" Mar 13 10:31:12 crc kubenswrapper[4930]: I0313 10:31:12.573017 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-6scqh" podUID="eb475308-399c-4168-8e73-5816fb8c30e9" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:12 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:12 crc kubenswrapper[4930]: > Mar 13 10:31:12 crc kubenswrapper[4930]: I0313 10:31:12.692911 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:31:13 crc kubenswrapper[4930]: I0313 10:31:13.986078 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" path="/var/lib/kubelet/pods/2b4f2c08-b04f-40e6-a90c-256225efe861/volumes" Mar 13 10:31:16 crc kubenswrapper[4930]: E0313 10:31:16.939592 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-f264ce8447bd7cb99d07f9e0d7f5c5f77f8ec22eceeb83c50151cb43ca78710f.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-conmon-f264ce8447bd7cb99d07f9e0d7f5c5f77f8ec22eceeb83c50151cb43ca78710f.scope\": RecentStats: unable to find data in memory cache]" Mar 13 10:31:17 crc kubenswrapper[4930]: I0313 10:31:17.113293 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/2.log" Mar 13 10:31:17 crc kubenswrapper[4930]: I0313 10:31:17.132784 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 13 10:31:17 crc kubenswrapper[4930]: I0313 10:31:17.143656 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 10:31:17 crc kubenswrapper[4930]: I0313 10:31:17.143712 4930 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f264ce8447bd7cb99d07f9e0d7f5c5f77f8ec22eceeb83c50151cb43ca78710f" exitCode=137 Mar 13 10:31:17 crc kubenswrapper[4930]: I0313 10:31:17.143764 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f264ce8447bd7cb99d07f9e0d7f5c5f77f8ec22eceeb83c50151cb43ca78710f"} Mar 13 10:31:17 crc kubenswrapper[4930]: I0313 10:31:17.143800 4930 scope.go:117] "RemoveContainer" containerID="4a1fbef37d7349b1be77763d642a29b52e831bc1c84f43d9151467d8e12a0db8" Mar 13 10:31:17 crc kubenswrapper[4930]: I0313 10:31:17.413024 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/prometheus-k8s-0" Mar 13 10:31:17 crc kubenswrapper[4930]: I0313 10:31:17.489329 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/prometheus-k8s-0" Mar 13 10:31:17 crc kubenswrapper[4930]: I0313 10:31:17.613273 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:31:18 crc kubenswrapper[4930]: I0313 10:31:18.156096 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/2.log" Mar 13 10:31:18 crc kubenswrapper[4930]: I0313 10:31:18.157645 4930 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 13 10:31:18 crc kubenswrapper[4930]: I0313 10:31:18.157769 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"830c9788f8fb2baaf8ea337458057e1b01ebce45cbdb3cb44c67fbb393af9f09"} Mar 13 10:31:18 crc kubenswrapper[4930]: I0313 10:31:18.199040 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/prometheus-k8s-0" Mar 13 10:31:19 crc kubenswrapper[4930]: I0313 10:31:19.320274 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:31:19 crc kubenswrapper[4930]: I0313 10:31:19.689568 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 13 10:31:20 crc kubenswrapper[4930]: I0313 10:31:20.833843 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jskr2" podUID="72b25617-0c24-4567-9134-4f9783e97387" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:20 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:20 crc kubenswrapper[4930]: > Mar 13 10:31:21 crc kubenswrapper[4930]: I0313 10:31:21.021618 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:21 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:21 crc kubenswrapper[4930]: > Mar 13 10:31:21 crc kubenswrapper[4930]: I0313 10:31:21.427153 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:31:21 crc kubenswrapper[4930]: I0313 10:31:21.481678 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:31:22 crc kubenswrapper[4930]: I0313 10:31:22.612167 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:31:23 crc kubenswrapper[4930]: E0313 10:31:23.102566 4930 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.238:57456->38.102.83.238:33987: write tcp 38.102.83.238:57456->38.102.83.238:33987: write: broken pipe Mar 13 10:31:23 crc kubenswrapper[4930]: I0313 10:31:23.126973 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 10:31:23 crc kubenswrapper[4930]: I0313 10:31:23.133464 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-zw9cp" Mar 13 10:31:26 crc kubenswrapper[4930]: I0313 10:31:26.153838 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6scqh"] Mar 13 10:31:26 crc kubenswrapper[4930]: I0313 10:31:26.154634 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6scqh" podUID="eb475308-399c-4168-8e73-5816fb8c30e9" containerName="registry-server" containerID="cri-o://0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc" gracePeriod=2 Mar 13 10:31:26 crc kubenswrapper[4930]: I0313 10:31:26.672936 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:31:26 crc kubenswrapper[4930]: I0313 10:31:26.673002 4930 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 13 10:31:26 crc kubenswrapper[4930]: I0313 10:31:26.673050 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.278379 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.278374 4930 generic.go:334] "Generic (PLEG): container finished" podID="eb475308-399c-4168-8e73-5816fb8c30e9" containerID="0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc" exitCode=0 Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.278407 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6scqh" event={"ID":"eb475308-399c-4168-8e73-5816fb8c30e9","Type":"ContainerDied","Data":"0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc"} Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.278995 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6scqh" event={"ID":"eb475308-399c-4168-8e73-5816fb8c30e9","Type":"ContainerDied","Data":"01302c574110474b85e7f0931cf9a33ecb1191147a06b1ee311f82d17a53df5b"} Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.279037 4930 scope.go:117] "RemoveContainer" containerID="0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.332297 4930 scope.go:117] "RemoveContainer" containerID="8689a8419f2ca8c4d4600a4c395c189c53e509edf1e59c056976d7231230ddf3" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.374828 4930 scope.go:117] "RemoveContainer" containerID="c7e7b0b13d783101548245c3d6eeb92023f88fdc0afe1d3745cbd5f7177b4725" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.383907 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb475308-399c-4168-8e73-5816fb8c30e9-catalog-content\") pod \"eb475308-399c-4168-8e73-5816fb8c30e9\" (UID: \"eb475308-399c-4168-8e73-5816fb8c30e9\") " Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.384056 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb475308-399c-4168-8e73-5816fb8c30e9-utilities\") pod \"eb475308-399c-4168-8e73-5816fb8c30e9\" (UID: \"eb475308-399c-4168-8e73-5816fb8c30e9\") " Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.384768 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb475308-399c-4168-8e73-5816fb8c30e9-utilities" (OuterVolumeSpecName: "utilities") pod "eb475308-399c-4168-8e73-5816fb8c30e9" (UID: "eb475308-399c-4168-8e73-5816fb8c30e9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.384906 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz8zg\" (UniqueName: \"kubernetes.io/projected/eb475308-399c-4168-8e73-5816fb8c30e9-kube-api-access-nz8zg\") pod \"eb475308-399c-4168-8e73-5816fb8c30e9\" (UID: \"eb475308-399c-4168-8e73-5816fb8c30e9\") " Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.387016 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/eb475308-399c-4168-8e73-5816fb8c30e9-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.405293 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb475308-399c-4168-8e73-5816fb8c30e9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "eb475308-399c-4168-8e73-5816fb8c30e9" (UID: "eb475308-399c-4168-8e73-5816fb8c30e9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.406938 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb475308-399c-4168-8e73-5816fb8c30e9-kube-api-access-nz8zg" (OuterVolumeSpecName: "kube-api-access-nz8zg") pod "eb475308-399c-4168-8e73-5816fb8c30e9" (UID: "eb475308-399c-4168-8e73-5816fb8c30e9"). InnerVolumeSpecName "kube-api-access-nz8zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.429107 4930 scope.go:117] "RemoveContainer" containerID="0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc" Mar 13 10:31:27 crc kubenswrapper[4930]: E0313 10:31:27.430694 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc\": container with ID starting with 0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc not found: ID does not exist" containerID="0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.430741 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc"} err="failed to get container status \"0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc\": rpc error: code = NotFound desc = could not find container \"0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc\": container with ID starting with 0064421db660856ee2eef2f442e7ebcd4c06ca0e4b1c95c67934d190177505bc not found: ID does not exist" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.430768 4930 scope.go:117] "RemoveContainer" containerID="8689a8419f2ca8c4d4600a4c395c189c53e509edf1e59c056976d7231230ddf3" Mar 13 10:31:27 crc kubenswrapper[4930]: E0313 10:31:27.431343 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8689a8419f2ca8c4d4600a4c395c189c53e509edf1e59c056976d7231230ddf3\": container with ID starting with 8689a8419f2ca8c4d4600a4c395c189c53e509edf1e59c056976d7231230ddf3 not found: ID does not exist" containerID="8689a8419f2ca8c4d4600a4c395c189c53e509edf1e59c056976d7231230ddf3" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.431515 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8689a8419f2ca8c4d4600a4c395c189c53e509edf1e59c056976d7231230ddf3"} err="failed to get container status \"8689a8419f2ca8c4d4600a4c395c189c53e509edf1e59c056976d7231230ddf3\": rpc error: code = NotFound desc = could not find container \"8689a8419f2ca8c4d4600a4c395c189c53e509edf1e59c056976d7231230ddf3\": container with ID starting with 8689a8419f2ca8c4d4600a4c395c189c53e509edf1e59c056976d7231230ddf3 not found: ID does not exist" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.431649 4930 scope.go:117] "RemoveContainer" containerID="c7e7b0b13d783101548245c3d6eeb92023f88fdc0afe1d3745cbd5f7177b4725" Mar 13 10:31:27 crc kubenswrapper[4930]: E0313 10:31:27.432193 4930 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7e7b0b13d783101548245c3d6eeb92023f88fdc0afe1d3745cbd5f7177b4725\": container with ID starting with c7e7b0b13d783101548245c3d6eeb92023f88fdc0afe1d3745cbd5f7177b4725 not found: ID does not exist" containerID="c7e7b0b13d783101548245c3d6eeb92023f88fdc0afe1d3745cbd5f7177b4725" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.432232 4930 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7e7b0b13d783101548245c3d6eeb92023f88fdc0afe1d3745cbd5f7177b4725"} err="failed to get container status \"c7e7b0b13d783101548245c3d6eeb92023f88fdc0afe1d3745cbd5f7177b4725\": rpc error: code = NotFound desc = could not find container \"c7e7b0b13d783101548245c3d6eeb92023f88fdc0afe1d3745cbd5f7177b4725\": container with ID starting with c7e7b0b13d783101548245c3d6eeb92023f88fdc0afe1d3745cbd5f7177b4725 not found: ID does not exist" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.489041 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz8zg\" (UniqueName: \"kubernetes.io/projected/eb475308-399c-4168-8e73-5816fb8c30e9-kube-api-access-nz8zg\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.489100 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/eb475308-399c-4168-8e73-5816fb8c30e9-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:27 crc kubenswrapper[4930]: I0313 10:31:27.641970 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:31:28 crc kubenswrapper[4930]: I0313 10:31:28.293620 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6scqh" Mar 13 10:31:28 crc kubenswrapper[4930]: I0313 10:31:28.343875 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6scqh"] Mar 13 10:31:28 crc kubenswrapper[4930]: I0313 10:31:28.360464 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6scqh"] Mar 13 10:31:29 crc kubenswrapper[4930]: I0313 10:31:29.848941 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:31:29 crc kubenswrapper[4930]: I0313 10:31:29.899218 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:31:29 crc kubenswrapper[4930]: I0313 10:31:29.983974 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb475308-399c-4168-8e73-5816fb8c30e9" path="/var/lib/kubelet/pods/eb475308-399c-4168-8e73-5816fb8c30e9/volumes" Mar 13 10:31:31 crc kubenswrapper[4930]: I0313 10:31:31.024450 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:31 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:31 crc kubenswrapper[4930]: > Mar 13 10:31:32 crc kubenswrapper[4930]: I0313 10:31:32.631853 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:31:35 crc kubenswrapper[4930]: I0313 10:31:35.377135 4930 generic.go:334] "Generic (PLEG): container finished" podID="4f1e0b58-9978-423f-84bc-f974dc6347b4" containerID="4e38434dc61e84fdfbc5b69f7d4692030e6412a93f7dbf0acb6c8116570ea79c" exitCode=1 Mar 13 10:31:35 crc kubenswrapper[4930]: I0313 10:31:35.377204 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" event={"ID":"4f1e0b58-9978-423f-84bc-f974dc6347b4","Type":"ContainerDied","Data":"4e38434dc61e84fdfbc5b69f7d4692030e6412a93f7dbf0acb6c8116570ea79c"} Mar 13 10:31:35 crc kubenswrapper[4930]: I0313 10:31:35.378838 4930 scope.go:117] "RemoveContainer" containerID="4e38434dc61e84fdfbc5b69f7d4692030e6412a93f7dbf0acb6c8116570ea79c" Mar 13 10:31:35 crc kubenswrapper[4930]: I0313 10:31:35.753256 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jskr2"] Mar 13 10:31:35 crc kubenswrapper[4930]: I0313 10:31:35.753870 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jskr2" podUID="72b25617-0c24-4567-9134-4f9783e97387" containerName="registry-server" containerID="cri-o://d30bd19fdd7ac708a5d36391a1303d851c5945faf75716fa6af1aab0e3373d8d" gracePeriod=2 Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.392282 4930 generic.go:334] "Generic (PLEG): container finished" podID="72b25617-0c24-4567-9134-4f9783e97387" containerID="d30bd19fdd7ac708a5d36391a1303d851c5945faf75716fa6af1aab0e3373d8d" exitCode=0 Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.392643 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jskr2" event={"ID":"72b25617-0c24-4567-9134-4f9783e97387","Type":"ContainerDied","Data":"d30bd19fdd7ac708a5d36391a1303d851c5945faf75716fa6af1aab0e3373d8d"} Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.397002 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-2k7k2" event={"ID":"4f1e0b58-9978-423f-84bc-f974dc6347b4","Type":"ContainerStarted","Data":"9ac17eda858752c42c4ea40783689437ca762843c6a2bbb7121543d02a497315"} Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.553673 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.631307 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b25617-0c24-4567-9134-4f9783e97387-catalog-content\") pod \"72b25617-0c24-4567-9134-4f9783e97387\" (UID: \"72b25617-0c24-4567-9134-4f9783e97387\") " Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.631827 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b25617-0c24-4567-9134-4f9783e97387-utilities\") pod \"72b25617-0c24-4567-9134-4f9783e97387\" (UID: \"72b25617-0c24-4567-9134-4f9783e97387\") " Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.631942 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8ng8l\" (UniqueName: \"kubernetes.io/projected/72b25617-0c24-4567-9134-4f9783e97387-kube-api-access-8ng8l\") pod \"72b25617-0c24-4567-9134-4f9783e97387\" (UID: \"72b25617-0c24-4567-9134-4f9783e97387\") " Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.632615 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72b25617-0c24-4567-9134-4f9783e97387-utilities" (OuterVolumeSpecName: "utilities") pod "72b25617-0c24-4567-9134-4f9783e97387" (UID: "72b25617-0c24-4567-9134-4f9783e97387"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.650737 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/72b25617-0c24-4567-9134-4f9783e97387-kube-api-access-8ng8l" (OuterVolumeSpecName: "kube-api-access-8ng8l") pod "72b25617-0c24-4567-9134-4f9783e97387" (UID: "72b25617-0c24-4567-9134-4f9783e97387"). InnerVolumeSpecName "kube-api-access-8ng8l". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.677377 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.681991 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.739480 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8ng8l\" (UniqueName: \"kubernetes.io/projected/72b25617-0c24-4567-9134-4f9783e97387-kube-api-access-8ng8l\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.739527 4930 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/72b25617-0c24-4567-9134-4f9783e97387-utilities\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.776495 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/72b25617-0c24-4567-9134-4f9783e97387-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "72b25617-0c24-4567-9134-4f9783e97387" (UID: "72b25617-0c24-4567-9134-4f9783e97387"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:31:36 crc kubenswrapper[4930]: I0313 10:31:36.842615 4930 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/72b25617-0c24-4567-9134-4f9783e97387-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:37 crc kubenswrapper[4930]: I0313 10:31:37.408838 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jskr2" event={"ID":"72b25617-0c24-4567-9134-4f9783e97387","Type":"ContainerDied","Data":"5fec651218e07599af9744e048a5122bf2d46adc98d1b8bbc1500c771d9e4f1e"} Mar 13 10:31:37 crc kubenswrapper[4930]: I0313 10:31:37.408942 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jskr2" Mar 13 10:31:37 crc kubenswrapper[4930]: I0313 10:31:37.409265 4930 scope.go:117] "RemoveContainer" containerID="d30bd19fdd7ac708a5d36391a1303d851c5945faf75716fa6af1aab0e3373d8d" Mar 13 10:31:37 crc kubenswrapper[4930]: I0313 10:31:37.443699 4930 scope.go:117] "RemoveContainer" containerID="d73f0e053d6b5ce889711fd5831e57cc9521cf198beb5dbefa12e2847b536c4d" Mar 13 10:31:37 crc kubenswrapper[4930]: I0313 10:31:37.479606 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jskr2"] Mar 13 10:31:37 crc kubenswrapper[4930]: I0313 10:31:37.482544 4930 scope.go:117] "RemoveContainer" containerID="fd7af99e4702111b803095fc9b71326c0cb2b2676d42d2dd0a4125e737600f5b" Mar 13 10:31:37 crc kubenswrapper[4930]: I0313 10:31:37.497302 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jskr2"] Mar 13 10:31:37 crc kubenswrapper[4930]: I0313 10:31:37.612908 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:31:37 crc kubenswrapper[4930]: I0313 10:31:37.985410 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="72b25617-0c24-4567-9134-4f9783e97387" path="/var/lib/kubelet/pods/72b25617-0c24-4567-9134-4f9783e97387/volumes" Mar 13 10:31:41 crc kubenswrapper[4930]: I0313 10:31:41.022174 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-568x4" podUID="4dd00f07-858c-42be-a09e-746f709bbcdb" containerName="registry-server" probeResult="failure" output=< Mar 13 10:31:41 crc kubenswrapper[4930]: timeout: failed to connect service ":50051" within 1s Mar 13 10:31:41 crc kubenswrapper[4930]: > Mar 13 10:31:42 crc kubenswrapper[4930]: I0313 10:31:42.615500 4930 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="7365e082-8383-4b59-974c-dfdedb789f77" containerName="cinder-scheduler" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.239758 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556630-rp4jp"] Mar 13 10:31:44 crc kubenswrapper[4930]: E0313 10:31:44.242902 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb475308-399c-4168-8e73-5816fb8c30e9" containerName="registry-server" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.243296 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb475308-399c-4168-8e73-5816fb8c30e9" containerName="registry-server" Mar 13 10:31:44 crc kubenswrapper[4930]: E0313 10:31:44.243400 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerName="extract-content" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.243482 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerName="extract-content" Mar 13 10:31:44 crc kubenswrapper[4930]: E0313 10:31:44.243550 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb475308-399c-4168-8e73-5816fb8c30e9" containerName="extract-utilities" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.243607 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb475308-399c-4168-8e73-5816fb8c30e9" containerName="extract-utilities" Mar 13 10:31:44 crc kubenswrapper[4930]: E0313 10:31:44.243667 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerName="extract-utilities" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.243721 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerName="extract-utilities" Mar 13 10:31:44 crc kubenswrapper[4930]: E0313 10:31:44.243808 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerName="registry-server" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.243880 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerName="registry-server" Mar 13 10:31:44 crc kubenswrapper[4930]: E0313 10:31:44.243989 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b25617-0c24-4567-9134-4f9783e97387" containerName="registry-server" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.244066 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b25617-0c24-4567-9134-4f9783e97387" containerName="registry-server" Mar 13 10:31:44 crc kubenswrapper[4930]: E0313 10:31:44.244167 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b25617-0c24-4567-9134-4f9783e97387" containerName="extract-utilities" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.244234 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b25617-0c24-4567-9134-4f9783e97387" containerName="extract-utilities" Mar 13 10:31:44 crc kubenswrapper[4930]: E0313 10:31:44.244311 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="72b25617-0c24-4567-9134-4f9783e97387" containerName="extract-content" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.244370 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="72b25617-0c24-4567-9134-4f9783e97387" containerName="extract-content" Mar 13 10:31:44 crc kubenswrapper[4930]: E0313 10:31:44.244450 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb475308-399c-4168-8e73-5816fb8c30e9" containerName="extract-content" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.244519 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb475308-399c-4168-8e73-5816fb8c30e9" containerName="extract-content" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.244888 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb475308-399c-4168-8e73-5816fb8c30e9" containerName="registry-server" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.244970 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="72b25617-0c24-4567-9134-4f9783e97387" containerName="registry-server" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.245035 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b4f2c08-b04f-40e6-a90c-256225efe861" containerName="registry-server" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.250996 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556630-rp4jp" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.263031 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6"] Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.266211 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.267761 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.268247 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.268872 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.280578 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.280880 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.283330 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556630-rp4jp"] Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.307613 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6"] Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.328554 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zxrx8\" (UniqueName: \"kubernetes.io/projected/b6f48b3e-7121-4e53-9c1f-11f06afd228a-kube-api-access-zxrx8\") pod \"collect-profiles-29556630-m24s6\" (UID: \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.328769 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6f48b3e-7121-4e53-9c1f-11f06afd228a-secret-volume\") pod \"collect-profiles-29556630-m24s6\" (UID: \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.328885 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4858g\" (UniqueName: \"kubernetes.io/projected/f9200cc5-d88a-483c-8963-18bff543ee19-kube-api-access-4858g\") pod \"auto-csr-approver-29556630-rp4jp\" (UID: \"f9200cc5-d88a-483c-8963-18bff543ee19\") " pod="openshift-infra/auto-csr-approver-29556630-rp4jp" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.329305 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6f48b3e-7121-4e53-9c1f-11f06afd228a-config-volume\") pod \"collect-profiles-29556630-m24s6\" (UID: \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.361606 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.361962 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-notification-agent" containerID="cri-o://7ca789ad26da2a5f016d61206574d5af655b682285e87a2263695e67d2fbfa1a" gracePeriod=30 Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.363811 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-central-agent" containerID="cri-o://a7e1727f351e4f00388a3572c09ac857860d977688466021a0018423f9e40cc4" gracePeriod=30 Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.363876 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="proxy-httpd" containerID="cri-o://3708622dae94ba1bab4db1b77a117f2a29cbd9c1db81d1c43a5ce2a5c8617e2d" gracePeriod=30 Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.363928 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="sg-core" containerID="cri-o://9c26120b518f3a76ed8b75222d28fee55d500545914ccd519b65a1dd5d5c3b14" gracePeriod=30 Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.435042 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zxrx8\" (UniqueName: \"kubernetes.io/projected/b6f48b3e-7121-4e53-9c1f-11f06afd228a-kube-api-access-zxrx8\") pod \"collect-profiles-29556630-m24s6\" (UID: \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.435604 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6f48b3e-7121-4e53-9c1f-11f06afd228a-secret-volume\") pod \"collect-profiles-29556630-m24s6\" (UID: \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.435692 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4858g\" (UniqueName: \"kubernetes.io/projected/f9200cc5-d88a-483c-8963-18bff543ee19-kube-api-access-4858g\") pod \"auto-csr-approver-29556630-rp4jp\" (UID: \"f9200cc5-d88a-483c-8963-18bff543ee19\") " pod="openshift-infra/auto-csr-approver-29556630-rp4jp" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.435897 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6f48b3e-7121-4e53-9c1f-11f06afd228a-config-volume\") pod \"collect-profiles-29556630-m24s6\" (UID: \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.437694 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6f48b3e-7121-4e53-9c1f-11f06afd228a-config-volume\") pod \"collect-profiles-29556630-m24s6\" (UID: \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.566827 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4858g\" (UniqueName: \"kubernetes.io/projected/f9200cc5-d88a-483c-8963-18bff543ee19-kube-api-access-4858g\") pod \"auto-csr-approver-29556630-rp4jp\" (UID: \"f9200cc5-d88a-483c-8963-18bff543ee19\") " pod="openshift-infra/auto-csr-approver-29556630-rp4jp" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.567211 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6f48b3e-7121-4e53-9c1f-11f06afd228a-secret-volume\") pod \"collect-profiles-29556630-m24s6\" (UID: \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.569136 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zxrx8\" (UniqueName: \"kubernetes.io/projected/b6f48b3e-7121-4e53-9c1f-11f06afd228a-kube-api-access-zxrx8\") pod \"collect-profiles-29556630-m24s6\" (UID: \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.605229 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556630-rp4jp" Mar 13 10:31:44 crc kubenswrapper[4930]: I0313 10:31:44.614324 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:45 crc kubenswrapper[4930]: E0313 10:31:45.333580 4930 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.238:45010->38.102.83.238:33987: write tcp 38.102.83.238:45010->38.102.83.238:33987: write: broken pipe Mar 13 10:31:45 crc kubenswrapper[4930]: I0313 10:31:45.537032 4930 generic.go:334] "Generic (PLEG): container finished" podID="d4640978-a913-419b-9faa-8b230f5e51f2" containerID="a7e1727f351e4f00388a3572c09ac857860d977688466021a0018423f9e40cc4" exitCode=0 Mar 13 10:31:45 crc kubenswrapper[4930]: I0313 10:31:45.537353 4930 generic.go:334] "Generic (PLEG): container finished" podID="d4640978-a913-419b-9faa-8b230f5e51f2" containerID="9c26120b518f3a76ed8b75222d28fee55d500545914ccd519b65a1dd5d5c3b14" exitCode=2 Mar 13 10:31:45 crc kubenswrapper[4930]: I0313 10:31:45.537364 4930 generic.go:334] "Generic (PLEG): container finished" podID="d4640978-a913-419b-9faa-8b230f5e51f2" containerID="7ca789ad26da2a5f016d61206574d5af655b682285e87a2263695e67d2fbfa1a" exitCode=0 Mar 13 10:31:45 crc kubenswrapper[4930]: I0313 10:31:45.537387 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4640978-a913-419b-9faa-8b230f5e51f2","Type":"ContainerDied","Data":"a7e1727f351e4f00388a3572c09ac857860d977688466021a0018423f9e40cc4"} Mar 13 10:31:45 crc kubenswrapper[4930]: I0313 10:31:45.537416 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4640978-a913-419b-9faa-8b230f5e51f2","Type":"ContainerDied","Data":"9c26120b518f3a76ed8b75222d28fee55d500545914ccd519b65a1dd5d5c3b14"} Mar 13 10:31:45 crc kubenswrapper[4930]: I0313 10:31:45.537425 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4640978-a913-419b-9faa-8b230f5e51f2","Type":"ContainerDied","Data":"7ca789ad26da2a5f016d61206574d5af655b682285e87a2263695e67d2fbfa1a"} Mar 13 10:31:45 crc kubenswrapper[4930]: I0313 10:31:45.537459 4930 scope.go:117] "RemoveContainer" containerID="1cd5e3669981d79173656007193a66857ed29e6d11519fa9bc1873410a276c0f" Mar 13 10:31:45 crc kubenswrapper[4930]: I0313 10:31:45.680287 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6"] Mar 13 10:31:45 crc kubenswrapper[4930]: I0313 10:31:45.733595 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556630-rp4jp"] Mar 13 10:31:45 crc kubenswrapper[4930]: I0313 10:31:45.769588 4930 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 13 10:31:46 crc kubenswrapper[4930]: I0313 10:31:46.554782 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" event={"ID":"b6f48b3e-7121-4e53-9c1f-11f06afd228a","Type":"ContainerStarted","Data":"7e7dc8f1fdee07aee2bd96fac47cc478e920ecc5bce848de8c7d077e963c947d"} Mar 13 10:31:46 crc kubenswrapper[4930]: I0313 10:31:46.555398 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" event={"ID":"b6f48b3e-7121-4e53-9c1f-11f06afd228a","Type":"ContainerStarted","Data":"9b58df76a8a657ca14fab2cc8a96f47d7096b03b4bd7cfaf0b3ec2fe53d01f47"} Mar 13 10:31:46 crc kubenswrapper[4930]: I0313 10:31:46.562520 4930 generic.go:334] "Generic (PLEG): container finished" podID="d4640978-a913-419b-9faa-8b230f5e51f2" containerID="3708622dae94ba1bab4db1b77a117f2a29cbd9c1db81d1c43a5ce2a5c8617e2d" exitCode=0 Mar 13 10:31:46 crc kubenswrapper[4930]: I0313 10:31:46.562611 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4640978-a913-419b-9faa-8b230f5e51f2","Type":"ContainerDied","Data":"3708622dae94ba1bab4db1b77a117f2a29cbd9c1db81d1c43a5ce2a5c8617e2d"} Mar 13 10:31:46 crc kubenswrapper[4930]: I0313 10:31:46.564328 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556630-rp4jp" event={"ID":"f9200cc5-d88a-483c-8963-18bff543ee19","Type":"ContainerStarted","Data":"0438fdef3852d842af6ead98bc3cc8d64376ff435dd73e4289a80eff97c87fea"} Mar 13 10:31:46 crc kubenswrapper[4930]: I0313 10:31:46.589625 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" podStartSLOduration=2.589604425 podStartE2EDuration="2.589604425s" podCreationTimestamp="2026-03-13 10:31:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-13 10:31:46.57946062 +0000 UTC m=+4747.329375297" watchObservedRunningTime="2026-03-13 10:31:46.589604425 +0000 UTC m=+4747.339519102" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.102671 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.162255 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.262198 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.336312 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-sg-core-conf-yaml\") pod \"d4640978-a913-419b-9faa-8b230f5e51f2\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.336424 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-combined-ca-bundle\") pod \"d4640978-a913-419b-9faa-8b230f5e51f2\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.336554 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4640978-a913-419b-9faa-8b230f5e51f2-log-httpd\") pod \"d4640978-a913-419b-9faa-8b230f5e51f2\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.336576 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-config-data\") pod \"d4640978-a913-419b-9faa-8b230f5e51f2\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.336606 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snkkd\" (UniqueName: \"kubernetes.io/projected/d4640978-a913-419b-9faa-8b230f5e51f2-kube-api-access-snkkd\") pod \"d4640978-a913-419b-9faa-8b230f5e51f2\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.336987 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-ceilometer-tls-certs\") pod \"d4640978-a913-419b-9faa-8b230f5e51f2\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.337033 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-scripts\") pod \"d4640978-a913-419b-9faa-8b230f5e51f2\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.337189 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4640978-a913-419b-9faa-8b230f5e51f2-run-httpd\") pod \"d4640978-a913-419b-9faa-8b230f5e51f2\" (UID: \"d4640978-a913-419b-9faa-8b230f5e51f2\") " Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.337883 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4640978-a913-419b-9faa-8b230f5e51f2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d4640978-a913-419b-9faa-8b230f5e51f2" (UID: "d4640978-a913-419b-9faa-8b230f5e51f2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.338091 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d4640978-a913-419b-9faa-8b230f5e51f2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d4640978-a913-419b-9faa-8b230f5e51f2" (UID: "d4640978-a913-419b-9faa-8b230f5e51f2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.346037 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4640978-a913-419b-9faa-8b230f5e51f2-kube-api-access-snkkd" (OuterVolumeSpecName: "kube-api-access-snkkd") pod "d4640978-a913-419b-9faa-8b230f5e51f2" (UID: "d4640978-a913-419b-9faa-8b230f5e51f2"). InnerVolumeSpecName "kube-api-access-snkkd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.354578 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-scripts" (OuterVolumeSpecName: "scripts") pod "d4640978-a913-419b-9faa-8b230f5e51f2" (UID: "d4640978-a913-419b-9faa-8b230f5e51f2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.439770 4930 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4640978-a913-419b-9faa-8b230f5e51f2-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.439805 4930 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d4640978-a913-419b-9faa-8b230f5e51f2-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.439815 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snkkd\" (UniqueName: \"kubernetes.io/projected/d4640978-a913-419b-9faa-8b230f5e51f2-kube-api-access-snkkd\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.439827 4930 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-scripts\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.584973 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d4640978-a913-419b-9faa-8b230f5e51f2" (UID: "d4640978-a913-419b-9faa-8b230f5e51f2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.610483 4930 generic.go:334] "Generic (PLEG): container finished" podID="b6f48b3e-7121-4e53-9c1f-11f06afd228a" containerID="7e7dc8f1fdee07aee2bd96fac47cc478e920ecc5bce848de8c7d077e963c947d" exitCode=0 Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.610569 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" event={"ID":"b6f48b3e-7121-4e53-9c1f-11f06afd228a","Type":"ContainerDied","Data":"7e7dc8f1fdee07aee2bd96fac47cc478e920ecc5bce848de8c7d077e963c947d"} Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.643923 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.644191 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d4640978-a913-419b-9faa-8b230f5e51f2","Type":"ContainerDied","Data":"2f4dbf9dcdf28ac45b6b206d225da1af5e1cfdfabc95f44458f1e62ed25c5cc8"} Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.644235 4930 scope.go:117] "RemoveContainer" containerID="a7e1727f351e4f00388a3572c09ac857860d977688466021a0018423f9e40cc4" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.645940 4930 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.662648 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4640978-a913-419b-9faa-8b230f5e51f2" (UID: "d4640978-a913-419b-9faa-8b230f5e51f2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.670258 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-config-data" (OuterVolumeSpecName: "config-data") pod "d4640978-a913-419b-9faa-8b230f5e51f2" (UID: "d4640978-a913-419b-9faa-8b230f5e51f2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.690591 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d4640978-a913-419b-9faa-8b230f5e51f2" (UID: "d4640978-a913-419b-9faa-8b230f5e51f2"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.747950 4930 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-config-data\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.747994 4930 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.748008 4930 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4640978-a913-419b-9faa-8b230f5e51f2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.758551 4930 scope.go:117] "RemoveContainer" containerID="3708622dae94ba1bab4db1b77a117f2a29cbd9c1db81d1c43a5ce2a5c8617e2d" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.843403 4930 scope.go:117] "RemoveContainer" containerID="9c26120b518f3a76ed8b75222d28fee55d500545914ccd519b65a1dd5d5c3b14" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.923819 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 13 10:31:47 crc kubenswrapper[4930]: I0313 10:31:47.986570 4930 scope.go:117] "RemoveContainer" containerID="7ca789ad26da2a5f016d61206574d5af655b682285e87a2263695e67d2fbfa1a" Mar 13 10:31:48 crc kubenswrapper[4930]: I0313 10:31:48.661939 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556630-rp4jp" event={"ID":"f9200cc5-d88a-483c-8963-18bff543ee19","Type":"ContainerStarted","Data":"8895c16c221e44e035dcae3428babc8a7940cc8505c5cc6306f9b5d1642fd9b2"} Mar 13 10:31:48 crc kubenswrapper[4930]: I0313 10:31:48.698116 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556630-rp4jp" podStartSLOduration=2.921526173 podStartE2EDuration="4.698090708s" podCreationTimestamp="2026-03-13 10:31:44 +0000 UTC" firstStartedPulling="2026-03-13 10:31:45.761064365 +0000 UTC m=+4746.510979042" lastFinishedPulling="2026-03-13 10:31:47.5376289 +0000 UTC m=+4748.287543577" observedRunningTime="2026-03-13 10:31:48.681082931 +0000 UTC m=+4749.430997608" watchObservedRunningTime="2026-03-13 10:31:48.698090708 +0000 UTC m=+4749.448005385" Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.224943 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.342522 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.364093 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.496039 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zxrx8\" (UniqueName: \"kubernetes.io/projected/b6f48b3e-7121-4e53-9c1f-11f06afd228a-kube-api-access-zxrx8\") pod \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\" (UID: \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\") " Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.496143 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6f48b3e-7121-4e53-9c1f-11f06afd228a-secret-volume\") pod \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\" (UID: \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\") " Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.496304 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6f48b3e-7121-4e53-9c1f-11f06afd228a-config-volume\") pod \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\" (UID: \"b6f48b3e-7121-4e53-9c1f-11f06afd228a\") " Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.497404 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6f48b3e-7121-4e53-9c1f-11f06afd228a-config-volume" (OuterVolumeSpecName: "config-volume") pod "b6f48b3e-7121-4e53-9c1f-11f06afd228a" (UID: "b6f48b3e-7121-4e53-9c1f-11f06afd228a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.498048 4930 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b6f48b3e-7121-4e53-9c1f-11f06afd228a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.524348 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6f48b3e-7121-4e53-9c1f-11f06afd228a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b6f48b3e-7121-4e53-9c1f-11f06afd228a" (UID: "b6f48b3e-7121-4e53-9c1f-11f06afd228a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.544651 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6f48b3e-7121-4e53-9c1f-11f06afd228a-kube-api-access-zxrx8" (OuterVolumeSpecName: "kube-api-access-zxrx8") pod "b6f48b3e-7121-4e53-9c1f-11f06afd228a" (UID: "b6f48b3e-7121-4e53-9c1f-11f06afd228a"). InnerVolumeSpecName "kube-api-access-zxrx8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.602210 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zxrx8\" (UniqueName: \"kubernetes.io/projected/b6f48b3e-7121-4e53-9c1f-11f06afd228a-kube-api-access-zxrx8\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.602250 4930 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b6f48b3e-7121-4e53-9c1f-11f06afd228a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.684641 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.684769 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29556630-m24s6" event={"ID":"b6f48b3e-7121-4e53-9c1f-11f06afd228a","Type":"ContainerDied","Data":"9b58df76a8a657ca14fab2cc8a96f47d7096b03b4bd7cfaf0b3ec2fe53d01f47"} Mar 13 10:31:49 crc kubenswrapper[4930]: I0313 10:31:49.684805 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b58df76a8a657ca14fab2cc8a96f47d7096b03b4bd7cfaf0b3ec2fe53d01f47" Mar 13 10:31:50 crc kubenswrapper[4930]: I0313 10:31:50.051834 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-568x4" Mar 13 10:31:50 crc kubenswrapper[4930]: I0313 10:31:50.161917 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-568x4" Mar 13 10:31:51 crc kubenswrapper[4930]: I0313 10:31:51.530650 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb"] Mar 13 10:31:51 crc kubenswrapper[4930]: I0313 10:31:51.547778 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29556585-5qpzb"] Mar 13 10:31:51 crc kubenswrapper[4930]: I0313 10:31:51.705168 4930 generic.go:334] "Generic (PLEG): container finished" podID="f9200cc5-d88a-483c-8963-18bff543ee19" containerID="8895c16c221e44e035dcae3428babc8a7940cc8505c5cc6306f9b5d1642fd9b2" exitCode=0 Mar 13 10:31:51 crc kubenswrapper[4930]: I0313 10:31:51.705212 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556630-rp4jp" event={"ID":"f9200cc5-d88a-483c-8963-18bff543ee19","Type":"ContainerDied","Data":"8895c16c221e44e035dcae3428babc8a7940cc8505c5cc6306f9b5d1642fd9b2"} Mar 13 10:31:51 crc kubenswrapper[4930]: I0313 10:31:51.983882 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed8cfe62-57c8-42c3-82bb-d712ff3f8265" path="/var/lib/kubelet/pods/ed8cfe62-57c8-42c3-82bb-d712ff3f8265/volumes" Mar 13 10:31:53 crc kubenswrapper[4930]: I0313 10:31:53.331203 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556630-rp4jp" Mar 13 10:31:53 crc kubenswrapper[4930]: I0313 10:31:53.411038 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4858g\" (UniqueName: \"kubernetes.io/projected/f9200cc5-d88a-483c-8963-18bff543ee19-kube-api-access-4858g\") pod \"f9200cc5-d88a-483c-8963-18bff543ee19\" (UID: \"f9200cc5-d88a-483c-8963-18bff543ee19\") " Mar 13 10:31:54 crc kubenswrapper[4930]: I0313 10:31:54.044388 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9200cc5-d88a-483c-8963-18bff543ee19-kube-api-access-4858g" (OuterVolumeSpecName: "kube-api-access-4858g") pod "f9200cc5-d88a-483c-8963-18bff543ee19" (UID: "f9200cc5-d88a-483c-8963-18bff543ee19"). InnerVolumeSpecName "kube-api-access-4858g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:31:54 crc kubenswrapper[4930]: I0313 10:31:54.081459 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556630-rp4jp" event={"ID":"f9200cc5-d88a-483c-8963-18bff543ee19","Type":"ContainerDied","Data":"0438fdef3852d842af6ead98bc3cc8d64376ff435dd73e4289a80eff97c87fea"} Mar 13 10:31:54 crc kubenswrapper[4930]: I0313 10:31:54.081507 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0438fdef3852d842af6ead98bc3cc8d64376ff435dd73e4289a80eff97c87fea" Mar 13 10:31:54 crc kubenswrapper[4930]: I0313 10:31:54.081591 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556630-rp4jp" Mar 13 10:31:54 crc kubenswrapper[4930]: I0313 10:31:54.144882 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4858g\" (UniqueName: \"kubernetes.io/projected/f9200cc5-d88a-483c-8963-18bff543ee19-kube-api-access-4858g\") on node \"crc\" DevicePath \"\"" Mar 13 10:31:54 crc kubenswrapper[4930]: I0313 10:31:54.201718 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556624-9rqsx"] Mar 13 10:31:54 crc kubenswrapper[4930]: I0313 10:31:54.230751 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556624-9rqsx"] Mar 13 10:31:55 crc kubenswrapper[4930]: I0313 10:31:55.983090 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ffa5b2b3-62be-4df6-a444-a1634dc993cb" path="/var/lib/kubelet/pods/ffa5b2b3-62be-4df6-a444-a1634dc993cb/volumes" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.137688 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556632-nf6vf"] Mar 13 10:32:00 crc kubenswrapper[4930]: E0313 10:32:00.138861 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="sg-core" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.138880 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="sg-core" Mar 13 10:32:00 crc kubenswrapper[4930]: E0313 10:32:00.138908 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9200cc5-d88a-483c-8963-18bff543ee19" containerName="oc" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.138916 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9200cc5-d88a-483c-8963-18bff543ee19" containerName="oc" Mar 13 10:32:00 crc kubenswrapper[4930]: E0313 10:32:00.138931 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-notification-agent" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.138939 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-notification-agent" Mar 13 10:32:00 crc kubenswrapper[4930]: E0313 10:32:00.138959 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="proxy-httpd" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.138966 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="proxy-httpd" Mar 13 10:32:00 crc kubenswrapper[4930]: E0313 10:32:00.139000 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6f48b3e-7121-4e53-9c1f-11f06afd228a" containerName="collect-profiles" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.139008 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6f48b3e-7121-4e53-9c1f-11f06afd228a" containerName="collect-profiles" Mar 13 10:32:00 crc kubenswrapper[4930]: E0313 10:32:00.139024 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-central-agent" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.139032 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-central-agent" Mar 13 10:32:00 crc kubenswrapper[4930]: E0313 10:32:00.139051 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-central-agent" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.139058 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-central-agent" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.139325 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="proxy-httpd" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.139351 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9200cc5-d88a-483c-8963-18bff543ee19" containerName="oc" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.139372 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6f48b3e-7121-4e53-9c1f-11f06afd228a" containerName="collect-profiles" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.139385 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-central-agent" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.139400 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="sg-core" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.139416 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-central-agent" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.139450 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" containerName="ceilometer-notification-agent" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.140260 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556632-nf6vf" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.147917 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.148240 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.148625 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.162444 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556632-nf6vf"] Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.296250 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7ncm\" (UniqueName: \"kubernetes.io/projected/559e5f10-fd0d-4084-afff-18db3c6bf111-kube-api-access-w7ncm\") pod \"auto-csr-approver-29556632-nf6vf\" (UID: \"559e5f10-fd0d-4084-afff-18db3c6bf111\") " pod="openshift-infra/auto-csr-approver-29556632-nf6vf" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.399834 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7ncm\" (UniqueName: \"kubernetes.io/projected/559e5f10-fd0d-4084-afff-18db3c6bf111-kube-api-access-w7ncm\") pod \"auto-csr-approver-29556632-nf6vf\" (UID: \"559e5f10-fd0d-4084-afff-18db3c6bf111\") " pod="openshift-infra/auto-csr-approver-29556632-nf6vf" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.423558 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7ncm\" (UniqueName: \"kubernetes.io/projected/559e5f10-fd0d-4084-afff-18db3c6bf111-kube-api-access-w7ncm\") pod \"auto-csr-approver-29556632-nf6vf\" (UID: \"559e5f10-fd0d-4084-afff-18db3c6bf111\") " pod="openshift-infra/auto-csr-approver-29556632-nf6vf" Mar 13 10:32:00 crc kubenswrapper[4930]: I0313 10:32:00.466696 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556632-nf6vf" Mar 13 10:32:01 crc kubenswrapper[4930]: I0313 10:32:01.013583 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556632-nf6vf"] Mar 13 10:32:01 crc kubenswrapper[4930]: I0313 10:32:01.158047 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556632-nf6vf" event={"ID":"559e5f10-fd0d-4084-afff-18db3c6bf111","Type":"ContainerStarted","Data":"2be2079f2d7794b84a02c9fb6ed7686da460f3b3374574e4921f1ff0ac7748c7"} Mar 13 10:32:03 crc kubenswrapper[4930]: I0313 10:32:03.214994 4930 generic.go:334] "Generic (PLEG): container finished" podID="559e5f10-fd0d-4084-afff-18db3c6bf111" containerID="8fd9adacbb4368545248d5eb470830dc22eab23796292217dd0050301eb065f0" exitCode=0 Mar 13 10:32:03 crc kubenswrapper[4930]: I0313 10:32:03.215511 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556632-nf6vf" event={"ID":"559e5f10-fd0d-4084-afff-18db3c6bf111","Type":"ContainerDied","Data":"8fd9adacbb4368545248d5eb470830dc22eab23796292217dd0050301eb065f0"} Mar 13 10:32:05 crc kubenswrapper[4930]: I0313 10:32:05.015770 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556632-nf6vf" Mar 13 10:32:05 crc kubenswrapper[4930]: I0313 10:32:05.136376 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7ncm\" (UniqueName: \"kubernetes.io/projected/559e5f10-fd0d-4084-afff-18db3c6bf111-kube-api-access-w7ncm\") pod \"559e5f10-fd0d-4084-afff-18db3c6bf111\" (UID: \"559e5f10-fd0d-4084-afff-18db3c6bf111\") " Mar 13 10:32:05 crc kubenswrapper[4930]: I0313 10:32:05.156725 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/559e5f10-fd0d-4084-afff-18db3c6bf111-kube-api-access-w7ncm" (OuterVolumeSpecName: "kube-api-access-w7ncm") pod "559e5f10-fd0d-4084-afff-18db3c6bf111" (UID: "559e5f10-fd0d-4084-afff-18db3c6bf111"). InnerVolumeSpecName "kube-api-access-w7ncm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:32:05 crc kubenswrapper[4930]: I0313 10:32:05.240239 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7ncm\" (UniqueName: \"kubernetes.io/projected/559e5f10-fd0d-4084-afff-18db3c6bf111-kube-api-access-w7ncm\") on node \"crc\" DevicePath \"\"" Mar 13 10:32:05 crc kubenswrapper[4930]: I0313 10:32:05.255056 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556632-nf6vf" event={"ID":"559e5f10-fd0d-4084-afff-18db3c6bf111","Type":"ContainerDied","Data":"2be2079f2d7794b84a02c9fb6ed7686da460f3b3374574e4921f1ff0ac7748c7"} Mar 13 10:32:05 crc kubenswrapper[4930]: I0313 10:32:05.255475 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2be2079f2d7794b84a02c9fb6ed7686da460f3b3374574e4921f1ff0ac7748c7" Mar 13 10:32:05 crc kubenswrapper[4930]: I0313 10:32:05.255401 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556632-nf6vf" Mar 13 10:32:06 crc kubenswrapper[4930]: I0313 10:32:06.102120 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556626-bfbhx"] Mar 13 10:32:06 crc kubenswrapper[4930]: I0313 10:32:06.116559 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556626-bfbhx"] Mar 13 10:32:07 crc kubenswrapper[4930]: I0313 10:32:07.985153 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44ba6b14-c48d-443a-bb78-cf3e560f3a46" path="/var/lib/kubelet/pods/44ba6b14-c48d-443a-bb78-cf3e560f3a46/volumes" Mar 13 10:32:12 crc kubenswrapper[4930]: I0313 10:32:12.308640 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:32:12 crc kubenswrapper[4930]: I0313 10:32:12.309214 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.078354 4930 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","podd4640978-a913-419b-9faa-8b230f5e51f2"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort podd4640978-a913-419b-9faa-8b230f5e51f2] : Timed out while waiting for systemd to remove kubepods-besteffort-podd4640978_a913_419b_9faa_8b230f5e51f2.slice" Mar 13 10:32:18 crc kubenswrapper[4930]: E0313 10:32:18.079160 4930 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort podd4640978-a913-419b-9faa-8b230f5e51f2] : unable to destroy cgroup paths for cgroup [kubepods besteffort podd4640978-a913-419b-9faa-8b230f5e51f2] : Timed out while waiting for systemd to remove kubepods-besteffort-podd4640978_a913_419b_9faa_8b230f5e51f2.slice" pod="openstack/ceilometer-0" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.403196 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.433267 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.448084 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.505517 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:32:18 crc kubenswrapper[4930]: E0313 10:32:18.507685 4930 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="559e5f10-fd0d-4084-afff-18db3c6bf111" containerName="oc" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.507727 4930 state_mem.go:107] "Deleted CPUSet assignment" podUID="559e5f10-fd0d-4084-afff-18db3c6bf111" containerName="oc" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.508802 4930 memory_manager.go:354] "RemoveStaleState removing state" podUID="559e5f10-fd0d-4084-afff-18db3c6bf111" containerName="oc" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.516800 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.533085 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.533105 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.533492 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.559213 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.691186 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.691256 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-scripts\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.691609 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.691665 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-run-httpd\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.691691 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6bcnz\" (UniqueName: \"kubernetes.io/projected/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-kube-api-access-6bcnz\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.691874 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-log-httpd\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.692008 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.692119 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-config-data\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.793287 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.793358 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-run-httpd\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.793394 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6bcnz\" (UniqueName: \"kubernetes.io/projected/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-kube-api-access-6bcnz\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.793473 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-log-httpd\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.793526 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.793581 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-config-data\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.793690 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.793719 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-scripts\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.794228 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-log-httpd\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.794487 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-run-httpd\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.805186 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.806642 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-scripts\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.813694 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.814622 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.814872 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-config-data\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.825645 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6bcnz\" (UniqueName: \"kubernetes.io/projected/b8588dfd-8c59-404f-a4bf-2d5bd91f8612-kube-api-access-6bcnz\") pod \"ceilometer-0\" (UID: \"b8588dfd-8c59-404f-a4bf-2d5bd91f8612\") " pod="openstack/ceilometer-0" Mar 13 10:32:18 crc kubenswrapper[4930]: I0313 10:32:18.870165 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 13 10:32:19 crc kubenswrapper[4930]: I0313 10:32:19.464100 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 13 10:32:20 crc kubenswrapper[4930]: I0313 10:32:20.007261 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4640978-a913-419b-9faa-8b230f5e51f2" path="/var/lib/kubelet/pods/d4640978-a913-419b-9faa-8b230f5e51f2/volumes" Mar 13 10:32:20 crc kubenswrapper[4930]: I0313 10:32:20.428075 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8588dfd-8c59-404f-a4bf-2d5bd91f8612","Type":"ContainerStarted","Data":"23b76672d1ae48c9e3b162fd5f08d4b71909c9c688b26f394dc3922ccc048ef3"} Mar 13 10:32:20 crc kubenswrapper[4930]: I0313 10:32:20.428396 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8588dfd-8c59-404f-a4bf-2d5bd91f8612","Type":"ContainerStarted","Data":"309d45c4e11169bc7c66dd22d0df9d23438a8c05f425c604f5bd6f757d2f7f37"} Mar 13 10:32:21 crc kubenswrapper[4930]: I0313 10:32:21.453832 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8588dfd-8c59-404f-a4bf-2d5bd91f8612","Type":"ContainerStarted","Data":"ba950cf60cee6f6e8f680342b00eef7e5f1cd92aa0f87495d752203395b19fea"} Mar 13 10:32:22 crc kubenswrapper[4930]: I0313 10:32:22.473362 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8588dfd-8c59-404f-a4bf-2d5bd91f8612","Type":"ContainerStarted","Data":"78f7b75b58a86d50ac91fada07e31aba1d68fd5a056c2a50673ef0ba79cd15ed"} Mar 13 10:32:25 crc kubenswrapper[4930]: I0313 10:32:25.506995 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b8588dfd-8c59-404f-a4bf-2d5bd91f8612","Type":"ContainerStarted","Data":"0a0d8728697434e8bc9e2ec7cfed35c75f44abe3dc1db70727b36df0b84c4aa0"} Mar 13 10:32:25 crc kubenswrapper[4930]: I0313 10:32:25.507584 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 13 10:32:25 crc kubenswrapper[4930]: I0313 10:32:25.534022 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.972013455 podStartE2EDuration="7.534004679s" podCreationTimestamp="2026-03-13 10:32:18 +0000 UTC" firstStartedPulling="2026-03-13 10:32:19.475241449 +0000 UTC m=+4780.225156126" lastFinishedPulling="2026-03-13 10:32:24.037232673 +0000 UTC m=+4784.787147350" observedRunningTime="2026-03-13 10:32:25.53325304 +0000 UTC m=+4786.283167717" watchObservedRunningTime="2026-03-13 10:32:25.534004679 +0000 UTC m=+4786.283919366" Mar 13 10:32:36 crc kubenswrapper[4930]: I0313 10:32:36.394648 4930 scope.go:117] "RemoveContainer" containerID="c62ea15f39c602bfe3561c09b828e926f13747cd5106b1bae48468bf120409a3" Mar 13 10:32:36 crc kubenswrapper[4930]: I0313 10:32:36.530035 4930 scope.go:117] "RemoveContainer" containerID="d55e708d8732d2a3a27eb859cbb1cbb1e26816bbf9ee88fb2731459c1b608416" Mar 13 10:32:36 crc kubenswrapper[4930]: I0313 10:32:36.583554 4930 scope.go:117] "RemoveContainer" containerID="7d6331a452b3dcbfaeaf0f6a4d7d4db4a044511b9751df063cffc1996d1d433c" Mar 13 10:32:42 crc kubenswrapper[4930]: I0313 10:32:42.308605 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:32:42 crc kubenswrapper[4930]: I0313 10:32:42.309178 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:32:48 crc kubenswrapper[4930]: E0313 10:32:48.287772 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4d5bf77_4258_4108_ad1b_f51fa45a3bab.slice/crio-5620f3a4a3a7c68ab03d8e44a6f4f2ed95ed87191aa5c9da5fa88541afd99535.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4d5bf77_4258_4108_ad1b_f51fa45a3bab.slice/crio-conmon-5620f3a4a3a7c68ab03d8e44a6f4f2ed95ed87191aa5c9da5fa88541afd99535.scope\": RecentStats: unable to find data in memory cache]" Mar 13 10:32:48 crc kubenswrapper[4930]: E0313 10:32:48.287795 4930 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4d5bf77_4258_4108_ad1b_f51fa45a3bab.slice/crio-5620f3a4a3a7c68ab03d8e44a6f4f2ed95ed87191aa5c9da5fa88541afd99535.scope\": RecentStats: unable to find data in memory cache]" Mar 13 10:32:48 crc kubenswrapper[4930]: I0313 10:32:48.815641 4930 generic.go:334] "Generic (PLEG): container finished" podID="e4d5bf77-4258-4108-ad1b-f51fa45a3bab" containerID="5620f3a4a3a7c68ab03d8e44a6f4f2ed95ed87191aa5c9da5fa88541afd99535" exitCode=0 Mar 13 10:32:48 crc kubenswrapper[4930]: I0313 10:32:48.815693 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" event={"ID":"e4d5bf77-4258-4108-ad1b-f51fa45a3bab","Type":"ContainerDied","Data":"5620f3a4a3a7c68ab03d8e44a6f4f2ed95ed87191aa5c9da5fa88541afd99535"} Mar 13 10:32:48 crc kubenswrapper[4930]: I0313 10:32:48.815951 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" event={"ID":"e4d5bf77-4258-4108-ad1b-f51fa45a3bab","Type":"ContainerStarted","Data":"13bb53ad0454102268c49e34b0f330ba888487c1e634ace29877366ebab5e736"} Mar 13 10:32:48 crc kubenswrapper[4930]: I0313 10:32:48.890425 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 13 10:32:56 crc kubenswrapper[4930]: I0313 10:32:56.517885 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 10:32:56 crc kubenswrapper[4930]: I0313 10:32:56.518459 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 10:33:12 crc kubenswrapper[4930]: I0313 10:33:12.308621 4930 patch_prober.go:28] interesting pod/machine-config-daemon-724mj container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 13 10:33:12 crc kubenswrapper[4930]: I0313 10:33:12.309151 4930 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 13 10:33:12 crc kubenswrapper[4930]: I0313 10:33:12.309198 4930 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-724mj" Mar 13 10:33:12 crc kubenswrapper[4930]: I0313 10:33:12.310800 4930 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2c05a807e57dc5147123981c1da8fc8655db8bfe897a8f8333698a2f312225b8"} pod="openshift-machine-config-operator/machine-config-daemon-724mj" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 13 10:33:12 crc kubenswrapper[4930]: I0313 10:33:12.311222 4930 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-724mj" podUID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerName="machine-config-daemon" containerID="cri-o://2c05a807e57dc5147123981c1da8fc8655db8bfe897a8f8333698a2f312225b8" gracePeriod=600 Mar 13 10:33:13 crc kubenswrapper[4930]: I0313 10:33:13.083269 4930 generic.go:334] "Generic (PLEG): container finished" podID="22188dce-43d2-4c7e-aa9b-7090a71eeb06" containerID="2c05a807e57dc5147123981c1da8fc8655db8bfe897a8f8333698a2f312225b8" exitCode=0 Mar 13 10:33:13 crc kubenswrapper[4930]: I0313 10:33:13.083354 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerDied","Data":"2c05a807e57dc5147123981c1da8fc8655db8bfe897a8f8333698a2f312225b8"} Mar 13 10:33:13 crc kubenswrapper[4930]: I0313 10:33:13.083930 4930 scope.go:117] "RemoveContainer" containerID="d76065217e6156a922fa7ccc9f364904f822f8eda997e939abd9b792619096db" Mar 13 10:33:14 crc kubenswrapper[4930]: I0313 10:33:14.098504 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-724mj" event={"ID":"22188dce-43d2-4c7e-aa9b-7090a71eeb06","Type":"ContainerStarted","Data":"db90101bc81a513ace6b2e6cf2ed2f44948b0a93246b1f9eb27db9f17008dbfb"} Mar 13 10:33:16 crc kubenswrapper[4930]: I0313 10:33:16.539211 4930 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 10:33:16 crc kubenswrapper[4930]: I0313 10:33:16.548237 4930 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-monitoring/metrics-server-647f59f4f4-vnqd8" Mar 13 10:34:00 crc kubenswrapper[4930]: I0313 10:34:00.174138 4930 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29556634-kr2bq"] Mar 13 10:34:00 crc kubenswrapper[4930]: I0313 10:34:00.176479 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556634-kr2bq" Mar 13 10:34:00 crc kubenswrapper[4930]: I0313 10:34:00.179660 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 13 10:34:00 crc kubenswrapper[4930]: I0313 10:34:00.179708 4930 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-55m8x" Mar 13 10:34:00 crc kubenswrapper[4930]: I0313 10:34:00.182289 4930 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 13 10:34:00 crc kubenswrapper[4930]: I0313 10:34:00.187406 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556634-kr2bq"] Mar 13 10:34:00 crc kubenswrapper[4930]: I0313 10:34:00.290729 4930 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jln2d\" (UniqueName: \"kubernetes.io/projected/3e7556c2-f175-477f-a32a-87ae6c0b3bcc-kube-api-access-jln2d\") pod \"auto-csr-approver-29556634-kr2bq\" (UID: \"3e7556c2-f175-477f-a32a-87ae6c0b3bcc\") " pod="openshift-infra/auto-csr-approver-29556634-kr2bq" Mar 13 10:34:00 crc kubenswrapper[4930]: I0313 10:34:00.394608 4930 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jln2d\" (UniqueName: \"kubernetes.io/projected/3e7556c2-f175-477f-a32a-87ae6c0b3bcc-kube-api-access-jln2d\") pod \"auto-csr-approver-29556634-kr2bq\" (UID: \"3e7556c2-f175-477f-a32a-87ae6c0b3bcc\") " pod="openshift-infra/auto-csr-approver-29556634-kr2bq" Mar 13 10:34:00 crc kubenswrapper[4930]: I0313 10:34:00.422953 4930 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jln2d\" (UniqueName: \"kubernetes.io/projected/3e7556c2-f175-477f-a32a-87ae6c0b3bcc-kube-api-access-jln2d\") pod \"auto-csr-approver-29556634-kr2bq\" (UID: \"3e7556c2-f175-477f-a32a-87ae6c0b3bcc\") " pod="openshift-infra/auto-csr-approver-29556634-kr2bq" Mar 13 10:34:00 crc kubenswrapper[4930]: I0313 10:34:00.510565 4930 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556634-kr2bq" Mar 13 10:34:01 crc kubenswrapper[4930]: I0313 10:34:01.885365 4930 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29556634-kr2bq"] Mar 13 10:34:02 crc kubenswrapper[4930]: I0313 10:34:02.722083 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556634-kr2bq" event={"ID":"3e7556c2-f175-477f-a32a-87ae6c0b3bcc","Type":"ContainerStarted","Data":"d3b2e0299b68ef00c1239a968e743a5693ebe09466cbf6ea01a0302becb90513"} Mar 13 10:34:06 crc kubenswrapper[4930]: I0313 10:34:06.817315 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556634-kr2bq" event={"ID":"3e7556c2-f175-477f-a32a-87ae6c0b3bcc","Type":"ContainerStarted","Data":"81decc972e799b44ba835a38971770b5d62a3b3d6fbd2b00ea285e25317992d4"} Mar 13 10:34:06 crc kubenswrapper[4930]: I0313 10:34:06.833303 4930 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29556634-kr2bq" podStartSLOduration=4.653022634 podStartE2EDuration="6.833283492s" podCreationTimestamp="2026-03-13 10:34:00 +0000 UTC" firstStartedPulling="2026-03-13 10:34:02.175371416 +0000 UTC m=+4882.925286093" lastFinishedPulling="2026-03-13 10:34:04.355632274 +0000 UTC m=+4885.105546951" observedRunningTime="2026-03-13 10:34:06.832542664 +0000 UTC m=+4887.582457351" watchObservedRunningTime="2026-03-13 10:34:06.833283492 +0000 UTC m=+4887.583198159" Mar 13 10:34:07 crc kubenswrapper[4930]: I0313 10:34:07.839107 4930 generic.go:334] "Generic (PLEG): container finished" podID="3e7556c2-f175-477f-a32a-87ae6c0b3bcc" containerID="81decc972e799b44ba835a38971770b5d62a3b3d6fbd2b00ea285e25317992d4" exitCode=0 Mar 13 10:34:07 crc kubenswrapper[4930]: I0313 10:34:07.839212 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556634-kr2bq" event={"ID":"3e7556c2-f175-477f-a32a-87ae6c0b3bcc","Type":"ContainerDied","Data":"81decc972e799b44ba835a38971770b5d62a3b3d6fbd2b00ea285e25317992d4"} Mar 13 10:34:09 crc kubenswrapper[4930]: I0313 10:34:09.370261 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556634-kr2bq" Mar 13 10:34:09 crc kubenswrapper[4930]: I0313 10:34:09.535896 4930 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jln2d\" (UniqueName: \"kubernetes.io/projected/3e7556c2-f175-477f-a32a-87ae6c0b3bcc-kube-api-access-jln2d\") pod \"3e7556c2-f175-477f-a32a-87ae6c0b3bcc\" (UID: \"3e7556c2-f175-477f-a32a-87ae6c0b3bcc\") " Mar 13 10:34:09 crc kubenswrapper[4930]: I0313 10:34:09.553842 4930 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e7556c2-f175-477f-a32a-87ae6c0b3bcc-kube-api-access-jln2d" (OuterVolumeSpecName: "kube-api-access-jln2d") pod "3e7556c2-f175-477f-a32a-87ae6c0b3bcc" (UID: "3e7556c2-f175-477f-a32a-87ae6c0b3bcc"). InnerVolumeSpecName "kube-api-access-jln2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 13 10:34:09 crc kubenswrapper[4930]: I0313 10:34:09.640212 4930 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jln2d\" (UniqueName: \"kubernetes.io/projected/3e7556c2-f175-477f-a32a-87ae6c0b3bcc-kube-api-access-jln2d\") on node \"crc\" DevicePath \"\"" Mar 13 10:34:09 crc kubenswrapper[4930]: I0313 10:34:09.865084 4930 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29556634-kr2bq" event={"ID":"3e7556c2-f175-477f-a32a-87ae6c0b3bcc","Type":"ContainerDied","Data":"d3b2e0299b68ef00c1239a968e743a5693ebe09466cbf6ea01a0302becb90513"} Mar 13 10:34:09 crc kubenswrapper[4930]: I0313 10:34:09.865224 4930 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29556634-kr2bq" Mar 13 10:34:09 crc kubenswrapper[4930]: I0313 10:34:09.865933 4930 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3b2e0299b68ef00c1239a968e743a5693ebe09466cbf6ea01a0302becb90513" Mar 13 10:34:09 crc kubenswrapper[4930]: I0313 10:34:09.936781 4930 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29556628-tjjtx"] Mar 13 10:34:09 crc kubenswrapper[4930]: I0313 10:34:09.945987 4930 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29556628-tjjtx"] Mar 13 10:34:09 crc kubenswrapper[4930]: I0313 10:34:09.987633 4930 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7acc7f58-2b75-428a-a2af-ced979a020c4" path="/var/lib/kubelet/pods/7acc7f58-2b75-428a-a2af-ced979a020c4/volumes" Mar 13 10:34:36 crc kubenswrapper[4930]: I0313 10:34:36.928961 4930 scope.go:117] "RemoveContainer" containerID="a6887e77eea5a880226dcc70437a06941c941f6c635f7833e650eb8a0d4580d6" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515154764170024457 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015154764171017375 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015154752105016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015154752105015462 5ustar corecore